[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/osism/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T00:02:12.862614Z",
                    "start": "2026-03-13T00:00:23.737641Z"
                },
                "id": "0242ac1f-6606-6d43-d34b-000000000002",
                "name": "Base pre"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:23.838692Z",
                            "start": "2026-03-13T00:00:23.790640Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000000c",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "14c/osism/14c2d5479cea4ebc868c9c5d14dcb57c"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000087",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:23.907721Z",
                            "start": "2026-03-13T00:00:23.863355Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000089",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.14\nJob: testbed-deploy-next-in-a-nutshell-with-tempest-ubuntu-24.04\nPipeline: periodic-midnight\nExecutor: 521e9411259a\nTriggered by: https://github.com/osism/testbed\nEvent ID: fcd5e110b62548aa83d02f3e7f3ac493\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:24.005374Z",
                            "start": "2026-03-13T00:00:23.926593Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000000d",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: orchestrator\nHostname: zuul-static-regiocloud-infra-1\nUsername: zuul-testbed03\nDistro: Debian 12.13\nProvider: static-testbed\nRegion: \nLabel: testbed-orchestrator\nProduct Name: OpenStack Nova\nInterface IP: 81.163.193.140\n",
                                    "zj_item": "orchestrator"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:24.246280Z",
                            "start": "2026-03-13T00:00:24.010779Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000000e",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:25.623354Z",
                            "start": "2026-03-13T00:00:24.255206Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000013",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "6fa9a7d157a21df8fa2454fd43a3074368b6b9d7",
                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "6fa9a7d157a21df8fa2454fd43a3074368b6b9d7",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360025.7393546-34-216277462693117/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "27df2cf18c3017537f3991579e8940d3",
                            "mode": "0644",
                            "owner": "root",
                            "size": 8194,
                            "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360025.7393546-34-216277462693117/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:27.769225Z",
                            "start": "2026-03-13T00:00:25.629307Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000014",
                        "name": "Copy ansible inventory to logs dir"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:28.480670Z",
                            "start": "2026-03-13T00:00:27.785208Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:28.541552Z",
                            "start": "2026-03-13T00:00:28.489546Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:01.855347",
                            "end": "2026-03-13 00:00:31.818465",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6d43-d34b-0000000000aa-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:00:29.963118",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa.pub\nThe key fingerprint is:\nSHA256:PzRaFc2h06GmNOk4NKmn/QyCvxI+k65+kcn0bkIv5vE zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|            .oo. |\n|         . . =o. |\n|        + + * .  |\n|    .  o = = .   |\n|   o +. S *      |\n|    B..+ * .     |\n|   oo*o + o      |\n|    @++. + .     |\n| .o*oBE.  o      |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:PzRaFc2h06GmNOk4NKmn/QyCvxI+k65+kcn0bkIv5vE zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|            .oo. |",
                                "|         . . =o. |",
                                "|        + + * .  |",
                                "|    .  o = = .   |",
                                "|   o +. S *      |",
                                "|    B..+ * .     |",
                                "|   oo*o + o      |",
                                "|    @++. + .     |",
                                "| .o*oBE.  o      |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6d43-d34b-0000000000aa-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:31.907728Z",
                            "start": "2026-03-13T00:00:28.611363Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000aa",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:31.956565Z",
                            "start": "2026-03-13T00:00:31.913351Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000ab",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:32.079329Z",
                            "start": "2026-03-13T00:00:32.023296Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000ca",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC9hZOK4zg7iDGqaFxQ9uy0ic4PLr3V1cgsb+AkMti0USqK3vbO/cXhafoyXykJhOxCykz2WfdbDKSFitk67gUPtm5U6DRz+wLcL4t9MZubLLDYqucs7UjBEzky8PUsAdyhZqp/caf+M+mep9zOhstPwVHwpfxXJFieFSGC32gOnkTeUSFJE1CxcdEFnFsd+ofs5hKIXNRxRFivaIxkqemyTix6oabXVeY3dn/Qw1blXMMbNyt94M1R1REx1nREXvX8gQNLUx4pDYNCRE+q7CgRkCNN4CSkP+g+ZhjAU80p0DyQh0Xje4du/EzBPEd8dMj9VvgrXLOg3qzVIXOvT9fLZUooH19iT7qvxQuj6FI393nssuR9/qUMmTTKIeb4YnY1Y8Mg7YQNO+eVTzbizPQiJSd+2Dk1K21/H/Md54lHS6ZhUKzbuap/nl62MwdWAbqMexqXl9Vlb1CW7yDWVEpDArPEwSWyOAsZkqcE3W3re8xlD1pqPAqtivXAgzqFYxM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul-testbed03",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC9hZOK4zg7iDGqaFxQ9uy0ic4PLr3V1cgsb+AkMti0USqK3vbO/cXhafoyXykJhOxCykz2WfdbDKSFitk67gUPtm5U6DRz+wLcL4t9MZubLLDYqucs7UjBEzky8PUsAdyhZqp/caf+M+mep9zOhstPwVHwpfxXJFieFSGC32gOnkTeUSFJE1CxcdEFnFsd+ofs5hKIXNRxRFivaIxkqemyTix6oabXVeY3dn/Qw1blXMMbNyt94M1R1REx1nREXvX8gQNLUx4pDYNCRE+q7CgRkCNN4CSkP+g+ZhjAU80p0DyQh0Xje4du/EzBPEd8dMj9VvgrXLOg3qzVIXOvT9fLZUooH19iT7qvxQuj6FI393nssuR9/qUMmTTKIeb4YnY1Y8Mg7YQNO+eVTzbizPQiJSd+2Dk1K21/H/Md54lHS6ZhUKzbuap/nl62MwdWAbqMexqXl9Vlb1CW7yDWVEpDArPEwSWyOAsZkqcE3W3re8xlD1pqPAqtivXAgzqFYxM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul-testbed03",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:33.176715Z",
                            "start": "2026-03-13T00:00:32.085544Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000cb",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed03/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed03/.ssh"
                                }
                            },
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed03/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul-testbed03",
                            "path": "/home/zuul-testbed03/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:33.534108Z",
                            "start": "2026-03-13T00:00:33.188069Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000cc",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed03/.ssh/id_rsa",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa",
                                "force": false,
                                "mode": 384,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa",
                                    "force": false,
                                    "mode": 384,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa"
                                },
                                "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa"
                            },
                            "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:33.981283Z",
                            "start": "2026-03-13T00:00:33.538928Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000cd",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed03/.ssh/id_rsa.pub",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa.pub",
                                "force": false,
                                "mode": 420,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa.pub",
                                    "force": false,
                                    "mode": 420,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa.pub"
                                },
                                "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa.pub"
                            },
                            "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa.pub"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:34.473748Z",
                            "start": "2026-03-13T00:00:33.991125Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000ce",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:34.529587Z",
                            "start": "2026-03-13T00:00:34.483846Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000ac",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/var/ssh/nodepool"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-0000000000b1",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:35.610312Z",
                            "start": "2026-03-13T00:00:34.535145Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000b3",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa"
                            ],
                            "delta": "0:00:00.064241",
                            "end": "2026-03-13 00:00:36.707506",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6d43-d34b-0000000000b9-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:00:36.643265",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/14c2d5479cea4ebc868c9c5d14dcb57c_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6d43-d34b-0000000000b9-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:36.796784Z",
                            "start": "2026-03-13T00:00:35.633949Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000b9",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:37.280683Z",
                            "start": "2026-03-13T00:00:36.802390Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000ba",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:37.325924Z",
                            "start": "2026-03-13T00:00:37.285430Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-0000000000bb",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001b",
                        "name": "start-zuul-console",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/start-zuul-console"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:37.818126Z",
                            "start": "2026-03-13T00:00:37.389993Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_info_dir": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:37.930287Z",
                            "start": "2026-03-13T00:00:37.842154Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000021",
                        "name": "Define zuul_info_dir fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:38.914348Z",
                            "start": "2026-03-13T00:00:37.941426Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000022",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "setup",
                            "ansible_facts": {
                                "ansible_all_ipv4_addresses": [
                                    "10.128.0.21"
                                ],
                                "ansible_all_ipv6_addresses": [
                                    "2a13:1a81:8000:2::33d",
                                    "fe80::f816:3eff:febf:60f0"
                                ],
                                "ansible_apparmor": {
                                    "status": "enabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "04/01/2014",
                                "ansible_bios_vendor": "SeaBIOS",
                                "ansible_bios_version": "1.15.0-1",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "QEMU",
                                "ansible_chassis_version": "pc-i440fx-6.2",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": "ttyS0,115200",
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_date_time": {
                                    "date": "2026-03-13",
                                    "day": "13",
                                    "epoch": "1773360040",
                                    "epoch_int": "1773360040",
                                    "hour": "00",
                                    "iso8601": "2026-03-13T00:00:40Z",
                                    "iso8601_basic": "20260313T000040603400",
                                    "iso8601_basic_short": "20260313T000040",
                                    "iso8601_micro": "2026-03-13T00:00:40.603400Z",
                                    "minute": "00",
                                    "month": "03",
                                    "second": "40",
                                    "time": "00:00:40",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Friday",
                                    "weekday_number": "5",
                                    "weeknumber": "10",
                                    "year": "2026"
                                },
                                "ansible_default_ipv4": {
                                    "address": "10.128.0.21",
                                    "alias": "ens3",
                                    "broadcast": "",
                                    "gateway": "10.128.0.17",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "netmask": "255.255.255.240",
                                    "network": "10.128.0.16",
                                    "prefix": "28",
                                    "type": "ether"
                                },
                                "ansible_default_ipv6": {
                                    "address": "2a13:1a81:8000:2::33d",
                                    "gateway": "fe80::f816:3eff:fe98:d9f",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "prefix": "128",
                                    "scope": "global",
                                    "type": "ether"
                                },
                                "ansible_device_links": {
                                    "ids": {
                                        "sda": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                        ],
                                        "sda1": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                        ],
                                        "sda14": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                        ],
                                        "sda15": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                        ]
                                    },
                                    "labels": {},
                                    "masters": {},
                                    "uuids": {
                                        "sda1": [
                                            "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                        ],
                                        "sda15": [
                                            "1E24-44C9"
                                        ]
                                    }
                                },
                                "ansible_devices": {
                                    "loop0": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop1": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop2": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop3": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop4": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop5": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop6": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop7": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "sda": {
                                        "holders": [],
                                        "host": "SCSI storage controller: Red Hat, Inc. Virtio SCSI",
                                        "links": {
                                            "ids": [
                                                "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                            ],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": "QEMU HARDDISK",
                                        "partitions": {
                                            "sda1": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                                    ]
                                                },
                                                "sectors": "209453023",
                                                "sectorsize": 512,
                                                "size": "99.87 GB",
                                                "start": "262144",
                                                "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                            },
                                            "sda14": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": []
                                                },
                                                "sectors": "6144",
                                                "sectorsize": 512,
                                                "size": "3.00 MB",
                                                "start": "2048",
                                                "uuid": null
                                            },
                                            "sda15": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "1E24-44C9"
                                                    ]
                                                },
                                                "sectors": "253952",
                                                "sectorsize": 512,
                                                "size": "124.00 MB",
                                                "start": "8192",
                                                "uuid": "1E24-44C9"
                                            }
                                        },
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "209715200",
                                        "sectorsize": "512",
                                        "size": "100.00 GB",
                                        "support_discard": "4096",
                                        "vendor": "QEMU",
                                        "virtual": 1
                                    }
                                },
                                "ansible_distribution": "Debian",
                                "ansible_distribution_file_parsed": true,
                                "ansible_distribution_file_path": "/etc/os-release",
                                "ansible_distribution_file_variety": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_minor_version": "13",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12.13",
                                "ansible_dns": {
                                    "nameservers": [
                                        "81.163.194.9",
                                        "81.163.194.10",
                                        "2a13:1a81::11:53",
                                        "2a13:1a81::12:53"
                                    ],
                                    "search": [
                                        "."
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 1004,
                                "ansible_effective_user_id": 1003,
                                "ansible_ens3": {
                                    "active": true,
                                    "device": "ens3",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "off [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "off [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "on",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "on [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "off [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "off [fixed]",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "on [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "off [fixed]",
                                        "tx_nocache_copy": "off",
                                        "tx_scatter_gather": "on",
                                        "tx_scatter_gather_fraglist": "off [fixed]",
                                        "tx_sctp_segmentation": "off [fixed]",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "off",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "off [fixed]",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "off [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "10.128.0.21",
                                        "broadcast": "",
                                        "netmask": "255.255.255.240",
                                        "network": "10.128.0.16",
                                        "prefix": "28"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "2a13:1a81:8000:2::33d",
                                            "prefix": "128",
                                            "scope": "global"
                                        },
                                        {
                                            "address": "fe80::f816:3eff:febf:60f0",
                                            "prefix": "64",
                                            "scope": "link"
                                        }
                                    ],
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "module": "virtio_net",
                                    "mtu": 1500,
                                    "pciid": "virtio1",
                                    "promisc": false,
                                    "speed": -1,
                                    "timestamping": [],
                                    "type": "ether"
                                },
                                "ansible_env": {
                                    "HOME": "/home/zuul-testbed03",
                                    "LANG": "C.UTF-8",
                                    "LOGNAME": "zuul-testbed03",
                                    "MOTD_SHOWN": "pam",
                                    "PATH": "/usr/local/bin:/usr/bin:/bin:/usr/games",
                                    "PWD": "/home/zuul-testbed03",
                                    "SHELL": "/bin/sh",
                                    "SSH_CLIENT": "81.163.193.69 44938 22",
                                    "SSH_CONNECTION": "81.163.193.69 44938 10.128.0.21 22",
                                    "USER": "zuul-testbed03",
                                    "XDG_RUNTIME_DIR": "/run/user/1003",
                                    "XDG_SESSION_CLASS": "user",
                                    "XDG_SESSION_ID": "22223",
                                    "XDG_SESSION_TYPE": "tty"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "Other",
                                "ansible_fqdn": "zuul-static-regiocloud-infra-1",
                                "ansible_hostname": "zuul-static-regiocloud-infra-1",
                                "ansible_hostnqn": "",
                                "ansible_interfaces": [
                                    "ens3",
                                    "lo"
                                ],
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "6.1.0-23-cloud-amd64",
                                "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15)",
                                "ansible_lo": {
                                    "active": true,
                                    "device": "lo",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "on [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "on [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "off [fixed]",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "off [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on [fixed]",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "on [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "on",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "off [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "on [fixed]",
                                        "tx_nocache_copy": "off [fixed]",
                                        "tx_scatter_gather": "on [fixed]",
                                        "tx_scatter_gather_fraglist": "on [fixed]",
                                        "tx_sctp_segmentation": "on",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "on",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "on",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "on [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "127.0.0.1",
                                        "broadcast": "",
                                        "netmask": "255.0.0.0",
                                        "network": "127.0.0.0",
                                        "prefix": "8"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "::1",
                                            "prefix": "128",
                                            "scope": "host"
                                        }
                                    ],
                                    "mtu": 65536,
                                    "promisc": false,
                                    "timestamping": [],
                                    "type": "loopback"
                                },
                                "ansible_loadavg": {
                                    "15m": 0.00927734375,
                                    "1m": 0.12353515625,
                                    "5m": 0.0302734375
                                },
                                "ansible_local": {},
                                "ansible_locally_reachable_ips": {
                                    "ipv4": [
                                        "10.128.0.21",
                                        "127.0.0.0/8",
                                        "127.0.0.1"
                                    ],
                                    "ipv6": [
                                        "::1",
                                        "2a13:1a81:8000:2::33d",
                                        "fe80::f816:3eff:febf:60f0"
                                    ]
                                },
                                "ansible_lsb": {
                                    "codename": "bookworm",
                                    "description": "Debian GNU/Linux 12 (bookworm)",
                                    "id": "Debian",
                                    "major_release": "12",
                                    "release": "12"
                                },
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_machine_id": "28f0fb0ddd3c48c38b3aa31603cb444d",
                                "ansible_memfree_mb": 532,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 7242,
                                        "used": 710
                                    },
                                    "real": {
                                        "free": 532,
                                        "total": 7952,
                                        "used": 7420
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 7952,
                                "ansible_mounts": [
                                    {
                                        "block_available": 17228699,
                                        "block_size": 4096,
                                        "block_total": 25759134,
                                        "block_used": 8530435,
                                        "device": "/dev/sda1",
                                        "fstype": "ext4",
                                        "inode_available": 5914091,
                                        "inode_total": 6545408,
                                        "inode_used": 631317,
                                        "mount": "/",
                                        "options": "rw,relatime,discard,errors=remount-ro",
                                        "size_available": 70568751104,
                                        "size_total": 105509412864,
                                        "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                    },
                                    {
                                        "block_available": 57419,
                                        "block_size": 2048,
                                        "block_total": 63339,
                                        "block_used": 5920,
                                        "device": "/dev/sda15",
                                        "fstype": "vfat",
                                        "inode_available": 0,
                                        "inode_total": 0,
                                        "inode_used": 0,
                                        "mount": "/boot/efi",
                                        "options": "rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=ascii,shortname=mixed,utf8,errors=remount-ro",
                                        "size_available": 117594112,
                                        "size_total": 129718272,
                                        "uuid": "1E24-44C9"
                                    }
                                ],
                                "ansible_nodename": "zuul-static-regiocloud-infra-1",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": [
                                        "tty0",
                                        "ttyS0,115200"
                                    ],
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_processor": [
                                    "0",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "1",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "2",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "3",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 4,
                                "ansible_processor_nproc": 4,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 4,
                                "ansible_product_name": "OpenStack Nova",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "26.2.1",
                                "ansible_python": {
                                    "executable": "/usr/bin/python3",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 2,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        2,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.2",
                                "ansible_real_group_id": 1004,
                                "ansible_real_user_id": 1003,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "systemd",
                                "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAOVugYBPOvEOPI9DA30PvhAEr88mJeaFmKN4pkfnBHV/1A1OlDKtX7lKWCBfEUilcd+tQFVKXm4KY+K9q4J5x5tpd7cL7LRDCFKN+py3+jc6U5J9EZ0o5cVG84z0c3XLEL6cCljG9Etx/0pXjjELRh1SGuUZweEx0143OSiKs/9DAAAAFQCT7R5MDTPEdFXQcZxA/fPRmtWFswAAAIBp8Iz2Njfcxfnh/mdxMztzkdNXulDt679+ghvl5588eCIDPa9xMtojivk53JaV6V4R+wsBN+MaKufzMOoVxuDkZcmw940H/1kcr3yOsJcZIkU1GRkJiNnSEIWJ6dVaXMh+wwbjw+8xNCA3+n3fARqUK906ICT5RsRrBe7W2K7dIAAAAIAL9qFr0vTidcXXhyWKN8I2iBsHSKfb6cjbseShw3ZeZE+kJ3IXz3z9oe0OwKp52GhvJrOXEkxST9YbtutbsVz5iS7lDpisvIWfFtvRRYaOtyvE8b4PWpHrk4s5tssqGDQ7y7zSqmRrj1sf5w0pwokPGmoSnPWUi0Vp0yhlysI/1w==",
                                "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss",
                                "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAuyO5R2SR4XT/Y65sItV7OQrUJXmGJ+//L9L/VugDohOWInFrkaZ6yFJMccVS7fmApekshaCGYnkIVdXV72hzg=",
                                "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256",
                                "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAILxLVCAcH99lPquKCT3PwHsvi+569sV28I73LxchEjeg",
                                "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519",
                                "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ3YX5Qhzq+HT8jTGGcX7rLsF3GCwObHFepfMmBZuD+gTfvdY5sYFSoXiT70rLPBjqETBTS0SQwmo8sd5AJI1vdgIqI5Sry8p52ng8tDg6C5CONt+xAaYOmE+ZDD/6naNX56b/gmgqBRQqPQHPmuptNxTlimdAcpoi+p9/UFBLTObMYmNL+5DTqw1+XECJ5j5kj5Ez40NT5w/1XnelEMcdw06WJZrzNEFt4Qpr2Zf+ybh0s+Xcn9u7d/gCbwXIPJvJnT0iSkiOzFvXJgOzwpvPtoi/5cV7g1LmgWsMLadT6gHof2poQaofC+1Bi0KiB2ANARMumvVBULx03qKvsqulsnlQAix2ffMraa3aMdUzK5aZo3hBx+H1uyEznUlyG9TUT5zhLIhNObetbpJgv6OJHiknBW6v/80FGD9t1pDiAwwBc/tJ8TyKpwyAdGMiEBIIVxTzx8QfPuApPJBJZXhMo2FBwRAEDEuuHbZu0znjhKKz4zlFT8TDrW+W7iThTvc=",
                                "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": [
                                    ""
                                ],
                                "ansible_system_capabilities_enforced": "True",
                                "ansible_system_vendor": "OpenStack Foundation",
                                "ansible_uptime_seconds": 48690015,
                                "ansible_user_dir": "/home/zuul-testbed03",
                                "ansible_user_gecos": "",
                                "ansible_user_gid": 1004,
                                "ansible_user_id": "zuul-testbed03",
                                "ansible_user_shell": "/bin/sh",
                                "ansible_user_uid": 1003,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "guest",
                                "ansible_virtualization_tech_guest": [
                                    "openstack"
                                ],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "openstack",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "fact_path": "/etc/ansible/facts.d",
                                    "filter": [],
                                    "gather_subset": [
                                        "all"
                                    ],
                                    "gather_timeout": 10
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:40.827601Z",
                            "start": "2026-03-13T00:00:38.922707Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000023",
                        "name": "Collect information about the host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "validate_host_hostname_sanitized": "orchestrator"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:40.971529Z",
                            "start": "2026-03-13T00:00:40.850240Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000024",
                        "name": "Sanitize hostname"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "a2a12eb79f07c9619827a6036fa5eb320859f5e6",
                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info/host-info.orchestrator.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "host-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "a2a12eb79f07c9619827a6036fa5eb320859f5e6",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info/host-info.orchestrator.yaml",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360041.11119-238-207016505686149/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "f8500a2de016aeccd6af6e03e6e261e9",
                            "mode": "0644",
                            "owner": "root",
                            "size": 19731,
                            "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360041.11119-238-207016505686149/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:42.549985Z",
                            "start": "2026-03-13T00:00:40.985684Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000025",
                        "name": "Write out all ansible variables/facts known for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_zuul_nolog_return": true,
                            "action": "zuul_debug_info",
                            "changed": false,
                            "df_i": "Filesystem      Inodes  IUsed   IFree IUse% Mounted on\nudev           1015018    312 1014706    1% /dev\ntmpfs          1017896    537 1017359    1% /run\n/dev/sda1      6545408 631317 5914091   10% /\ntmpfs          1017896      1 1017895    1% /dev/shm\ntmpfs          1017896      2 1017894    1% /run/lock\n/dev/sda15           0      0       0     - /boot/efi\ntmpfs           203579     17  203562    1% /run/user/1001\ntmpfs           203579     17  203562    1% /run/user/1006\ntmpfs           203579     17  203562    1% /run/user/1003\ntmpfs           203579     17  203562    1% /run/user/1004\n",
                            "df_m": "Filesystem     1M-blocks  Used Available Use% Mounted on\nudev                3965     0      3965   0% /dev\ntmpfs                796     1       795   1% /run\n/dev/sda1         100622 29196     67300  31% /\ntmpfs               3977     0      3977   0% /dev/shm\ntmpfs                  5     0         5   0% /run/lock\n/dev/sda15           124    12       113  10% /boot/efi\ntmpfs                796     0       796   0% /run/user/1001\ntmpfs                796     0       796   0% /run/user/1006\ntmpfs                796     0       796   0% /run/user/1003\ntmpfs                796     0       796   0% /run/user/1004\n",
                            "image_manifest_files": [],
                            "invocation": {
                                "module_args": {
                                    "image_manifest": null,
                                    "image_manifest_files": [
                                        "/etc/dib-builddate.txt",
                                        "/etc/image-hostname.txt"
                                    ],
                                    "ipv4_route_required": false,
                                    "ipv6_route_required": false,
                                    "traceroute_host": null
                                }
                            },
                            "network_interfaces": "1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n    inet 127.0.0.1/8 scope host lo\n       valid_lft forever preferred_lft forever\n    inet6 ::1/128 scope host noprefixroute \n       valid_lft forever preferred_lft forever\n2: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000\n    link/ether fa:16:3e:bf:60:f0 brd ff:ff:ff:ff:ff:ff\n    altname enp0s3\n    inet 10.128.0.21/28 metric 100 brd 10.128.0.31 scope global dynamic ens3\n       valid_lft 24825sec preferred_lft 24825sec\n    inet6 2a13:1a81:8000:2::33d/128 scope global noprefixroute \n       valid_lft forever preferred_lft forever\n    inet6 fe80::f816:3eff:febf:60f0/64 scope link \n       valid_lft forever preferred_lft forever\n",
                            "network_neighbors": "10.128.0.17 dev ens3 lladdr fa:16:3e:e3:0c:52 REACHABLE \nfe80::f816:3eff:fe98:d9f dev ens3 lladdr fa:16:3e:98:0d:9f router STALE \n",
                            "network_routing_v4": "default via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n10.128.0.16/28 dev ens3 proto kernel scope link src 10.128.0.21 metric 100 \n10.128.0.17 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n10.128.0.18 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n81.163.194.9 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n81.163.194.10 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n169.254.169.254 via 10.128.0.18 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n",
                            "network_routing_v6": "2a13:1a81:8000:2::300/120 dev ens3 proto ra metric 100 pref medium\nfe80::/64 dev ens3 proto kernel metric 256 pref medium\ndefault via fe80::f816:3eff:fe98:d9f dev ens3 proto ra metric 100 mtu 1500 pref medium\n",
                            "proc_cpuinfo": "processor\t: 0\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 0\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 0\ninitial apicid\t: 0\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 1\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 1\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 1\ninitial apicid\t: 1\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 2\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 2\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 2\ninitial apicid\t: 2\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 3\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 3\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 3\ninitial apicid\t: 3\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\n",
                            "traceroute": null,
                            "uname": "Linux zuul-static-regiocloud-infra-1 6.1.0-23-cloud-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:43.251747Z",
                            "start": "2026-03-13T00:00:42.554554Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000027",
                        "name": "Collect information about zuul worker"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "4cfb9db7b7cc52e75975b4ea96e30acf0004e1fd",
                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info/zuul-info.orchestrator.txt",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "zuul-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "4cfb9db7b7cc52e75975b4ea96e30acf0004e1fd",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/zuul-info/zuul-info.orchestrator.txt",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360043.3657076-277-77356139448055/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "0687bb0837dc9ca8b036130d42c45970",
                            "mode": "0644",
                            "owner": "root",
                            "size": 3369,
                            "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/.ansible/tmp/ansible-tmp-1773360043.3657076-277-77356139448055/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:44.605720Z",
                            "start": "2026-03-13T00:00:43.255859Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-000000000028",
                        "name": "Write out all zuul information for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:00:44.899107Z",
                            "start": "2026-03-13T00:00:44.613782Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000002c",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --compress --delete-after --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --out-format='<<CHANGED>>%i %n%L' /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/src zuul-testbed03@81.163.193.140:.",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": false,
                                    "delete": true,
                                    "dest": "zuul-testbed03@81.163.193.140:.",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": false,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "push",
                                    "owner": false,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": true,
                                    "rsync_opts": [],
                                    "rsync_path": null,
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/src",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": ".d..t...... src/\n.d..t...... src/github.com/\n.d..t...... src/github.com/osism/\n.d..t...... src/github.com/osism/ansible-collection-commons/\n<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-commons/.flake8\n<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-commons/README.md\n<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/description\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/index\n<f.st...... src/github.com/osism/ansible-collection-commons/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/20/b890adf3f31755b76bb35f6a85b471bc8611a0\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/\ncd+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/3f/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/3f/3853c41092b877696fd88dd0b0e64ad9ae191e\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/56/c6294e632c254549c6f8a8f6b2ec82a7ca9475\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/70/87df98cedc09d10112e545be17def164506b0f\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/\ncd+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/7e/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/7e/214843694d0dda9b3b5e599d2d1642d7fe94fe\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/b9/f3cb764ba2adef259d06ea75ca42878b99c111\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/e8/f562135696db3b0035ce27bfd6eda9519ec43a\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/\n<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/fb/f496297b7f2cdde5b294c741e04907868a06f7\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/\n<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/\n<f.st...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-services/\n<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-services/.flake8\n<f..t...... src/github.com/osism/ansible-collection-services/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-services/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-services/README.md\n<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-services/.git/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-services/.git/description\n<f..t...... src/github.com/osism/ansible-collection-services/.git/index\n<f.st...... src/github.com/osism/ansible-collection-services/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-services/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-services/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/09/9877719b9c4050e2c4194f238b68435c8d13f9\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/45c18a581110097708fa8077577946117184d5\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1e/fd16074cc6bea86831de22e4f1ad929170e1ef\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1f/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1f/7021b98c891b7a22079fc5ce63de4fbb23451a\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/baa696f26b3d28594fc5f336a58ad086119b51\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/28/fa13117bed5e1b25592f85da500c9b332ccb8c\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/2a/746d7728a201fefd9e184c48b9ba914e1b6084\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/30/77a397798a3b682c223ba02a62ac9541c21e80\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4e/937145d13b6360a376369f2c4c1b765099004d\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/72fc5c8da2fbab28c318a47fda85ed94a496d5\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/61/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/61/6c4bc24d6f449c1e15f37d01d6cfdd30a023cd\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/02fa927ebf5c80a14c6755646873712aa21705\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/68/52a3a42c25236015b8242dff834c9bc175c9cb\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6f/1a7fa6c5b315b8c64d576ec7a74a1fc8b9252f\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/75/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/75/98a6412096a227921bb9f11312fa1a1aaa44ff\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/7c/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/7c/166b96895d11bfff0947a1710552f256a682fe\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/44e0d3ddd44b0067ed30791a7f11fdb54dbbdf\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/e8931276538c4be6f59fafe29b8a1ff9f19df0\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/88/1d5461053b6f86f35c63de6bf89c2601be1b42\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8a/2ed47a86e336183bf5421866d177b8f7793621\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/7037cbc10715de30234839daecdb6d338db3a8\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/d796c8699443e29829130f7566d723fdb064c8\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/92/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/92/f595b15b5c44505c6375fc780fc8a7581c9f0a\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/95/a1a4c804cb2966bdb38ea0848b0beace188d5c\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/95/eade9661e9f5a2adadcad31f186fd464b69eec\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/9a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/9a/1290e868aa83d5df5a0b03c32e477f221a1680\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a0/e83416d624fb10a8eae3c94ae66a49b509c14f\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a4/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a4/68b55c5974bb0598142dcefdc8d03458b56310\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a8/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a8/ef214cc47e25f6f8e3e4e30e3073c767010688\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/aa/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/aa/a04863653a2361deb5b2bc3e6f982d0580281e\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ac/53de0b381cbe31ed0537e312688c03c3e7317d\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ba/9a4f0f9c8e0bb623f6a5f00ff5d62bb94a5ec1\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/bd/2deeb0500bea24f8cdaf0e918c5f1882045dd2\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/c9/fa05f95375f6281289163aac2d872e3efb568b\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/cf/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/cf/fb65c6bb455cd47e135ced1d99a1c28ac93393\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ee/fe454140e7de3a5ee0dc6c0b6c36322e7ee899\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ef/7deeb4579605284f54b655ce2adbf14594bd44\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/fa/47d8a04e45c4b19f035492bdc2e79db626758b\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.github/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-services/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-services/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2\n<f.st...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/\n<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml\n<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/\n<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2\n<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/\n<f.st...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2\n<f.st...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2\n.d..t...... src/github.com/osism/terraform-base/\n<f..t...... src/github.com/osism/terraform-base/.yamllint.yml\n<f..t...... src/github.com/osism/terraform-base/.zuul.yaml\n<f..t...... src/github.com/osism/terraform-base/LICENSE\n<f..t...... src/github.com/osism/terraform-base/README.md\n.d..t...... src/github.com/osism/terraform-base/.git/\n<f..t...... src/github.com/osism/terraform-base/.git/HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/config\n<f..t...... src/github.com/osism/terraform-base/.git/config.worktree\n<f..t...... src/github.com/osism/terraform-base/.git/description\n<f..t...... src/github.com/osism/terraform-base/.git/index\n<f..t...... src/github.com/osism/terraform-base/.git/packed-refs\n.d..t...... src/github.com/osism/terraform-base/.git/branches/\n.d..t...... src/github.com/osism/terraform-base/.git/hooks/\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample\n.d..t...... src/github.com/osism/terraform-base/.git/info/\n<f..t...... src/github.com/osism/terraform-base/.git/info/exclude\n.d..t...... src/github.com/osism/terraform-base/.git/logs/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/objects/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/00/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/01/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/02/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/05/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/06/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/07/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/08/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/10/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/12/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/15/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/16/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/17/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/20/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/21/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/24/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/26/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/29/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/2c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/36/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/37/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/41/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/43/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/46/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/47/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/48/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/49/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/52/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/55/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/56/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/58/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/5d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/61/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/63/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/66/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/67/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/68/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/6d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/70/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/72/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/76/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/80/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/81/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/82/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/83/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/84/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/85/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/86/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/88/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/90/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/91/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/92/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/95/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/96/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/98/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9e/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/aa/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/af/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ba/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ce/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/da/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/db/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dd/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/de/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/df/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ea/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ee/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ef/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/fb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ff/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/info/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/pack/\n.d..t...... src/github.com/osism/terraform-base/.git/refs/\n.d..t...... src/github.com/osism/terraform-base/.github/\n<f..t...... src/github.com/osism/terraform-base/.github/renovate.json\n.d..t...... src/github.com/osism/terraform-base/playbooks/\n<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml\n.d..t...... src/github.com/osism/terraform-base/testbed-default/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/environments/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars\n.d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf\n.d..t...... src/github.com/osism/testbed/\n<f..t...... src/github.com/osism/testbed/.ansible-lint\n<f..t...... src/github.com/osism/testbed/.flake8\n<f..t...... src/github.com/osism/testbed/.gitignore\n<f..t...... src/github.com/osism/testbed/.yamllint.yml\n<f..t...... src/github.com/osism/testbed/.zuul.yaml\n<f..t...... src/github.com/osism/testbed/LICENSE\n<f..t...... src/github.com/osism/testbed/Makefile\n<f..t...... src/github.com/osism/testbed/README.md\n<f..t...... src/github.com/osism/testbed/ansible.cfg\n<f..t...... src/github.com/osism/testbed/gilt.yml\n<f..t...... src/github.com/osism/testbed/requirements.txt\n<f..t...... src/github.com/osism/testbed/requirements.yml\n<f..t...... src/github.com/osism/testbed/tox.ini\n.d..t...... src/github.com/osism/testbed/.git/\n<f..t...... src/github.com/osism/testbed/.git/HEAD\n<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/testbed/.git/config\n<f..t...... src/github.com/osism/testbed/.git/config.worktree\n<f..t...... src/github.com/osism/testbed/.git/description\n<f..t...... src/github.com/osism/testbed/.git/index\n<f..t...... src/github.com/osism/testbed/.git/packed-refs\n.d..t...... src/github.com/osism/testbed/.git/branches/\n.d..t...... src/github.com/osism/testbed/.git/hooks/\n<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample\n.d..t...... src/github.com/osism/testbed/.git/info/\n<f..t...... src/github.com/osism/testbed/.git/info/exclude\n.d..t...... src/github.com/osism/testbed/.git/logs/\n<f..t...... src/github.com/osism/testbed/.git/logs/HEAD\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/testbed/.git/objects/\n.d..t...... src/github.com/osism/testbed/.git/objects/00/\n.d..t...... src/github.com/osism/testbed/.git/objects/01/\n.d..t...... src/github.com/osism/testbed/.git/objects/02/\n.d..t...... src/github.com/osism/testbed/.git/objects/03/\n.d..t...... src/github.com/osism/testbed/.git/objects/04/\n.d..t...... src/github.com/osism/testbed/.git/objects/05/\n.d..t...... src/github.com/osism/testbed/.git/objects/06/\n.d..t...... src/github.com/osism/testbed/.git/objects/07/\n.d..t...... src/github.com/osism/testbed/.git/objects/08/\n.d..t...... src/github.com/osism/testbed/.git/objects/09/\n.d..t...... src/github.com/osism/testbed/.git/objects/0a/\n.d..t...... src/github.com/osism/testbed/.git/objects/0b/\n.d..t...... src/github.com/osism/testbed/.git/objects/0c/\n.d..t...... src/github.com/osism/testbed/.git/objects/0d/\n.d..t...... src/github.com/osism/testbed/.git/objects/0e/\n.d..t...... src/github.com/osism/testbed/.git/objects/0f/\n.d..t...... src/github.com/osism/testbed/.git/objects/10/\n.d..t...... src/github.com/osism/testbed/.git/objects/11/\n.d..t...... src/github.com/osism/testbed/.git/objects/12/\n.d..t...... src/github.com/osism/testbed/.git/objects/13/\n.d..t...... src/github.com/osism/testbed/.git/objects/14/\n.d..t...... src/github.com/osism/testbed/.git/objects/15/\n.d..t...... src/github.com/osism/testbed/.git/objects/16/\n.d..t...... src/github.com/osism/testbed/.git/objects/17/\n.d..t...... src/github.com/osism/testbed/.git/objects/18/\n.d..t...... src/github.com/osism/testbed/.git/objects/19/\n.d..t...... src/github.com/osism/testbed/.git/objects/1a/\n.d..t...... src/github.com/osism/testbed/.git/objects/1b/\n.d..t...... src/github.com/osism/testbed/.git/objects/1c/\n.d..t...... src/github.com/osism/testbed/.git/objects/1d/\n.d..t...... src/github.com/osism/testbed/.git/objects/1e/\n.d..t...... src/github.com/osism/testbed/.git/objects/1f/\n.d..t...... src/github.com/osism/testbed/.git/objects/20/\n.d..t...... src/github.com/osism/testbed/.git/objects/21/\n.d..t...... src/github.com/osism/testbed/.git/objects/22/\n.d..t...... src/github.com/osism/testbed/.git/objects/23/\n.d..t...... src/github.com/osism/testbed/.git/objects/24/\n.d..t...... src/github.com/osism/testbed/.git/objects/25/\n.d..t...... src/github.com/osism/testbed/.git/objects/26/\n.d..t...... src/github.com/osism/testbed/.git/objects/27/\n.d..t...... src/github.com/osism/testbed/.git/objects/28/\n.d..t...... src/github.com/osism/testbed/.git/objects/29/\n.d..t...... src/github.com/osism/testbed/.git/objects/2a/\n.d..t...... src/github.com/osism/testbed/.git/objects/2b/\n.d..t...... src/github.com/osism/testbed/.git/objects/2c/\n.d..t...... src/github.com/osism/testbed/.git/objects/2d/\n.d..t...... src/github.com/osism/testbed/.git/objects/2e/\n.d..t...... src/github.com/osism/testbed/.git/objects/2f/\n.d..t...... src/github.com/osism/testbed/.git/objects/30/\n.d..t...... src/github.com/osism/testbed/.git/objects/31/\n.d..t...... src/github.com/osism/testbed/.git/objects/32/\n.d..t...... src/github.com/osism/testbed/.git/objects/33/\n.d..t...... src/github.com/osism/testbed/.git/objects/34/\n.d..t...... src/github.com/osism/testbed/.git/objects/35/\n.d..t...... src/github.com/osism/testbed/.git/objects/36/\n.d..t...... src/github.com/osism/testbed/.git/objects/37/\n.d..t...... src/github.com/osism/testbed/.git/objects/38/\n.d..t...... src/github.com/osism/testbed/.git/objects/39/\n.d..t...... src/github.com/osism/testbed/.git/objects/3a/\n.d..t...... src/github.com/osism/testbed/.git/objects/3b/\n.d..t...... src/github.com/osism/testbed/.git/objects/3c/\n.d..t...... src/github.com/osism/testbed/.git/objects/3d/\n.d..t...... src/github.com/osism/testbed/.git/objects/3e/\n.d..t...... src/github.com/osism/testbed/.git/objects/3f/\n.d..t...... src/github.com/osism/testbed/.git/objects/40/\n.d..t...... src/github.com/osism/testbed/.git/objects/41/\n.d..t...... src/github.com/osism/testbed/.git/objects/42/\n.d..t...... src/github.com/osism/testbed/.git/objects/43/\n.d..t...... src/github.com/osism/testbed/.git/objects/44/\n.d..t...... src/github.com/osism/testbed/.git/objects/45/\n.d..t...... src/github.com/osism/testbed/.git/objects/46/\n.d..t...... src/github.com/osism/testbed/.git/objects/47/\n.d..t...... src/github.com/osism/testbed/.git/objects/48/\n.d..t...... src/github.com/osism/testbed/.git/objects/49/\n.d..t...... src/github.com/osism/testbed/.git/objects/4a/\n.d..t...... src/github.com/osism/testbed/.git/objects/4c/\n.d..t...... src/github.com/osism/testbed/.git/objects/4d/\n.d..t...... src/github.com/osism/testbed/.git/objects/4e/\n.d..t...... src/github.com/osism/testbed/.git/objects/4f/\n.d..t...... src/github.com/osism/testbed/.git/objects/50/\n.d..t...... src/github.com/osism/testbed/.git/objects/51/\n.d..t...... src/github.com/osism/testbed/.git/objects/52/\n.d..t...... src/github.com/osism/testbed/.git/objects/53/\n.d..t...... src/github.com/osism/testbed/.git/objects/54/\n.d..t...... src/github.com/osism/testbed/.git/objects/55/\n.d..t...... src/github.com/osism/testbed/.git/objects/56/\n.d..t...... src/github.com/osism/testbed/.git/objects/57/\n.d..t...... src/github.com/osism/testbed/.git/objects/58/\n.d..t...... src/github.com/osism/testbed/.git/objects/59/\n.d..t...... src/github.com/osism/testbed/.git/objects/5a/\n.d..t...... src/github.com/osism/testbed/.git/objects/5b/\n.d..t...... src/github.com/osism/testbed/.git/objects/5c/\n.d..t...... src/github.com/osism/testbed/.git/objects/5d/\n.d..t...... src/github.com/osism/testbed/.git/objects/5e/\n.d..t...... src/github.com/osism/testbed/.git/objects/5f/\n.d..t...... src/github.com/osism/testbed/.git/objects/60/\n.d..t...... src/github.com/osism/testbed/.git/objects/61/\n.d..t...... src/github.com/osism/testbed/.git/objects/64/\n.d..t...... src/github.com/osism/testbed/.git/objects/65/\n.d..t...... src/github.com/osism/testbed/.git/objects/66/\n.d..t...... src/github.com/osism/testbed/.git/objects/67/\n.d..t...... src/github.com/osism/testbed/.git/objects/68/\n.d..t...... src/github.com/osism/testbed/.git/objects/69/\n.d..t...... src/github.com/osism/testbed/.git/objects/6a/\n.d..t...... src/github.com/osism/testbed/.git/objects/6b/\n.d..t...... src/github.com/osism/testbed/.git/objects/6c/\n.d..t...... src/github.com/osism/testbed/.git/objects/6d/\n.d..t...... src/github.com/osism/testbed/.git/objects/6e/\n.d..t...... src/github.com/osism/testbed/.git/objects/6f/\n.d..t...... src/github.com/osism/testbed/.git/objects/70/\n.d..t...... src/github.com/osism/testbed/.git/objects/71/\n.d..t...... src/github.com/osism/testbed/.git/objects/72/\n.d..t...... src/github.com/osism/testbed/.git/objects/73/\n.d..t...... src/github.com/osism/testbed/.git/objects/74/\n.d..t...... src/github.com/osism/testbed/.git/objects/75/\n.d..t...... src/github.com/osism/testbed/.git/objects/76/\n.d..t...... src/github.com/osism/testbed/.git/objects/77/\n.d..t...... src/github.com/osism/testbed/.git/objects/78/\n.d..t...... src/github.com/osism/testbed/.git/objects/79/\n.d..t...... src/github.com/osism/testbed/.git/objects/7a/\n.d..t...... src/github.com/osism/testbed/.git/objects/7b/\n.d..t...... src/github.com/osism/testbed/.git/objects/7c/\n.d..t...... src/github.com/osism/testbed/.git/objects/7d/\n.d..t...... src/github.com/osism/testbed/.git/objects/7e/\n.d..t...... src/github.com/osism/testbed/.git/objects/7f/\n.d..t...... src/github.com/osism/testbed/.git/objects/80/\n.d..t...... src/github.com/osism/testbed/.git/objects/81/\n.d..t...... src/github.com/osism/testbed/.git/objects/82/\n.d..t...... src/github.com/osism/testbed/.git/objects/83/\n.d..t...... src/github.com/osism/testbed/.git/objects/84/\n.d..t...... src/github.com/osism/testbed/.git/objects/85/\n.d..t...... src/github.com/osism/testbed/.git/objects/86/\n.d..t...... src/github.com/osism/testbed/.git/objects/87/\n.d..t...... src/github.com/osism/testbed/.git/objects/8a/\n.d..t...... src/github.com/osism/testbed/.git/objects/8b/\n.d..t...... src/github.com/osism/testbed/.git/objects/8c/\n.d..t...... src/github.com/osism/testbed/.git/objects/8e/\n.d..t...... src/github.com/osism/testbed/.git/objects/8f/\n.d..t...... src/github.com/osism/testbed/.git/objects/90/\n.d..t...... src/github.com/osism/testbed/.git/objects/91/\n.d..t...... src/github.com/osism/testbed/.git/objects/92/\n.d..t...... src/github.com/osism/testbed/.git/objects/93/\n.d..t...... src/github.com/osism/testbed/.git/objects/94/\n.d..t...... src/github.com/osism/testbed/.git/objects/95/\n.d..t...... src/github.com/osism/testbed/.git/objects/96/\n.d..t...... src/github.com/osism/testbed/.git/objects/97/\n.d..t...... src/github.com/osism/testbed/.git/objects/98/\n.d..t...... src/github.com/osism/testbed/.git/objects/9a/\n.d..t...... src/github.com/osism/testbed/.git/objects/9c/\n.d..t...... src/github.com/osism/testbed/.git/objects/9e/\n.d..t...... src/github.com/osism/testbed/.git/objects/9f/\n.d..t...... src/github.com/osism/testbed/.git/objects/a0/\n.d..t...... src/github.com/osism/testbed/.git/objects/a1/\n.d..t...... src/github.com/osism/testbed/.git/objects/a2/\n.d..t...... src/github.com/osism/testbed/.git/objects/a3/\n.d..t...... src/github.com/osism/testbed/.git/objects/a4/\n.d..t...... src/github.com/osism/testbed/.git/objects/a5/\n.d..t...... src/github.com/osism/testbed/.git/objects/a6/\n.d..t...... src/github.com/osism/testbed/.git/objects/a7/\n.d..t...... src/github.com/osism/testbed/.git/objects/a8/\n.d..t...... src/github.com/osism/testbed/.git/objects/a9/\n.d..t...... src/github.com/osism/testbed/.git/objects/aa/\n.d..t...... src/github.com/osism/testbed/.git/objects/ab/\n.d..t...... src/github.com/osism/testbed/.git/objects/ac/\n.d..t...... src/github.com/osism/testbed/.git/objects/ad/\n.d..t...... src/github.com/osism/testbed/.git/objects/ae/\n.d..t...... src/github.com/osism/testbed/.git/objects/af/\n.d..t...... src/github.com/osism/testbed/.git/objects/b0/\n.d..t...... src/github.com/osism/testbed/.git/objects/b1/\n.d..t...... src/github.com/osism/testbed/.git/objects/b2/\n.d..t...... src/github.com/osism/testbed/.git/objects/b3/\n.d..t...... src/github.com/osism/testbed/.git/objects/b4/\n.d..t...... src/github.com/osism/testbed/.git/objects/b5/\n.d..t...... src/github.com/osism/testbed/.git/objects/b6/\n.d..t...... src/github.com/osism/testbed/.git/objects/b7/\n.d..t...... src/github.com/osism/testbed/.git/objects/b8/\n.d..t...... src/github.com/osism/testbed/.git/objects/b9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ba/\n.d..t...... src/github.com/osism/testbed/.git/objects/bb/\n.d..t...... src/github.com/osism/testbed/.git/objects/bc/\n.d..t...... src/github.com/osism/testbed/.git/objects/bd/\n.d..t...... src/github.com/osism/testbed/.git/objects/be/\n.d..t...... src/github.com/osism/testbed/.git/objects/bf/\n.d..t...... src/github.com/osism/testbed/.git/objects/c0/\n.d..t...... src/github.com/osism/testbed/.git/objects/c1/\n.d..t...... src/github.com/osism/testbed/.git/objects/c2/\n.d..t...... src/github.com/osism/testbed/.git/objects/c3/\n.d..t...... src/github.com/osism/testbed/.git/objects/c4/\n.d..t...... src/github.com/osism/testbed/.git/objects/c5/\n.d..t...... src/github.com/osism/testbed/.git/objects/c6/\n.d..t...... src/github.com/osism/testbed/.git/objects/c7/\n.d..t...... src/github.com/osism/testbed/.git/objects/c8/\n.d..t...... src/github.com/osism/testbed/.git/objects/c9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ca/\n.d..t...... src/github.com/osism/testbed/.git/objects/cb/\n.d..t...... src/github.com/osism/testbed/.git/objects/cc/\n.d..t...... src/github.com/osism/testbed/.git/objects/cd/\n.d..t...... src/github.com/osism/testbed/.git/objects/ce/\n.d..t...... src/github.com/osism/testbed/.git/objects/cf/\n.d..t...... src/github.com/osism/testbed/.git/objects/d0/\n.d..t...... src/github.com/osism/testbed/.git/objects/d2/\n.d..t...... src/github.com/osism/testbed/.git/objects/d3/\n.d..t...... src/github.com/osism/testbed/.git/objects/d5/\n.d..t...... src/github.com/osism/testbed/.git/objects/d6/\n.d..t...... src/github.com/osism/testbed/.git/objects/d7/\n.d..t...... src/github.com/osism/testbed/.git/objects/d8/\n.d..t...... src/github.com/osism/testbed/.git/objects/d9/\n.d..t...... src/github.com/osism/testbed/.git/objects/da/\n.d..t...... src/github.com/osism/testbed/.git/objects/db/\n.d..t...... src/github.com/osism/testbed/.git/objects/dc/\n.d..t...... src/github.com/osism/testbed/.git/objects/dd/\n.d..t...... src/github.com/osism/testbed/.git/objects/de/\n.d..t...... src/github.com/osism/testbed/.git/objects/df/\n.d..t...... src/github.com/osism/testbed/.git/objects/e0/\n.d..t...... src/github.com/osism/testbed/.git/objects/e2/\n.d..t...... src/github.com/osism/testbed/.git/objects/e3/\n.d..t...... src/github.com/osism/testbed/.git/objects/e4/\n.d..t...... src/github.com/osism/testbed/.git/objects/e5/\n.d..t...... src/github.com/osism/testbed/.git/objects/e6/\n.d..t...... src/github.com/osism/testbed/.git/objects/e7/\n.d..t...... src/github.com/osism/testbed/.git/objects/e8/\n.d..t...... src/github.com/osism/testbed/.git/objects/e9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ea/\n.d..t...... src/github.com/osism/testbed/.git/objects/eb/\n.d..t...... src/github.com/osism/testbed/.git/objects/ec/\n.d..t...... src/github.com/osism/testbed/.git/objects/ed/\n.d..t...... src/github.com/osism/testbed/.git/objects/ee/\n.d..t...... src/github.com/osism/testbed/.git/objects/ef/\n.d..t...... src/github.com/osism/testbed/.git/objects/f0/\n.d..t...... src/github.com/osism/testbed/.git/objects/f1/\n.d..t...... src/github.com/osism/testbed/.git/objects/f2/\n.d..t...... src/github.com/osism/testbed/.git/objects/f3/\n.d..t...... src/github.com/osism/testbed/.git/objects/f4/\n.d..t...... src/github.com/osism/testbed/.git/objects/f5/\n.d..t...... src/github.com/osism/testbed/.git/objects/f6/\n.d..t...... src/github.com/osism/testbed/.git/objects/f8/\n.d..t...... src/github.com/osism/testbed/.git/objects/f9/\n.d..t...... src/github.com/osism/testbed/.git/objects/fa/\n.d..t...... src/github.com/osism/testbed/.git/objects/fb/\n.d..t...... src/github.com/osism/testbed/.git/objects/fc/\n.d..t...... src/github.com/osism/testbed/.git/objects/fd/\n.d..t...... src/github.com/osism/testbed/.git/objects/fe/\n.d..t...... src/github.com/osism/testbed/.git/objects/ff/\n.d..t...... src/github.com/osism/testbed/.git/objects/info/\n.d..t...... src/github.com/osism/testbed/.git/objects/pack/\n.d..t...... src/github.com/osism/testbed/.git/refs/\n.d..t...... src/github.com/osism/testbed/.github/\n<f..t...... src/github.com/osism/testbed/.github/renovate.json\n.d..t...... src/github.com/osism/testbed/.github/workflows/\n<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml\n.d..t...... src/github.com/osism/testbed/ansible/\n<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml\n<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml\n.d..t...... src/github.com/osism/testbed/containerlab/\n<f..t...... src/github.com/osism/testbed/containerlab/README.md\n<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml\n.d..t...... src/github.com/osism/testbed/containerlab/files/\n<f..t...... src/github.com/osism/testbed/containerlab/files/daemons\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf\n.d..t...... src/github.com/osism/testbed/contrib/\n<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh\n<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py\n<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql\n<f..t...... src/github.com/osism/testbed/contrib/semver2.sh\n<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py\n.d..t...... src/github.com/osism/testbed/contrib/ansible/\n<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml\n<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample\n.d..t...... src/github.com/osism/testbed/contrib/assets/\n<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot\n.d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster\n.d..t...... src/github.com/osism/testbed/contrib/openstack-cli/\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example\n.d..t...... src/github.com/osism/testbed/contrib/ownca/\n<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml\n.d..t...... src/github.com/osism/testbed/environments/\n<f..t...... src/github.com/osism/testbed/environments/.vault_pass\n<f..t...... src/github.com/osism/testbed/environments/ansible.cfg\n<f..t...... src/github.com/osism/testbed/environments/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/images.yml\n<f..t...... src/github.com/osism/testbed/environments/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph.test/\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph/\n<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/\n<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/files/\n<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact\n.d..t...... src/github.com/osism/testbed/environments/custom/templates/\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2\n.d..t...... src/github.com/osism/testbed/environments/generic/\n<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/images.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/\n<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/files/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/\n<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/hosts\n<f..t...... src/github.com/osism/testbed/environments/manager/images.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/files/\n<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml\n.d..t...... src/github.com/osism/testbed/environments/monitoring/\n<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/openstack/\n<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub\n<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/openrc\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst\n.d..t...... src/github.com/osism/testbed/inventory/\n<f..t...... src/github.com/osism/testbed/inventory/10-custom\n<f..t...... src/github.com/osism/testbed/inventory/20-roles\n<f..t...... src/github.com/osism/testbed/inventory/99-overwrite\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/all/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml\n.d..t...... src/github.com/osism/testbed/netbox/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml\n.d..t...... src/github.com/osism/testbed/netbox/moduletypes/\n<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep\n.d..t...... src/github.com/osism/testbed/netbox/resources/\n<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml\n.d..t...... src/github.com/osism/testbed/network/\n<f..t...... src/github.com/osism/testbed/network/iptables.sh\n<f..t...... src/github.com/osism/testbed/network/vxlan.sh\n.d..t...... src/github.com/osism/testbed/playbooks/\n<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml\n<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml\n<f..t...... src/github.com/osism/testbed/playbooks/post.yml\n<f..t...... src/github.com/osism/testbed/playbooks/pre.yml\n<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml\n.d..t...... src/github.com/osism/testbed/playbooks/templates/\n<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2\n.d..t...... src/github.com/osism/testbed/playbooks/vars/\n<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml\n<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml\n.d..t...... src/github.com/osism/testbed/scripts/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/check.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh\n<f..t...... src/github.com/osism/testbed/scripts/include.sh\n<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh\n<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh\n.d..t...... src/github.com/osism/testbed/scripts/bootstrap/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh\n.d..t...... src/github.com/osism/testbed/scripts/check/\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh\n.d..t...... src/github.com/osism/testbed/scripts/customisations/\n<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh\n.d..t...... src/github.com/osism/testbed/scripts/deploy/\n<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/scripts/upgrade/\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/terraform/\n<f..t...... src/github.com/osism/testbed/terraform/Makefile\n<f..t...... src/github.com/osism/testbed/terraform/README.md\n<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample\n<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/local.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample\n.d..t...... src/github.com/osism/testbed/terraform/environments/\n<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars\n.d..t...... src/github.com/osism/testbed/terraform/scripts/\n<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/\n*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/environment\n*deleting   src/github.com/osism/testbed/terraform/.terraform/\n*deleting   src/github.com/osism/testbed/terraform/variables.tf\n*deleting   src/github.com/osism/testbed/terraform/provider.tf\n*deleting   src/github.com/osism/testbed/terraform/outputs.tf\n*deleting   src/github.com/osism/testbed/terraform/nova.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron.tf\n*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/manager.tf\n*deleting   src/github.com/osism/testbed/terraform/main.tf\n*deleting   src/github.com/osism/testbed/terraform/local.env\n*deleting   src/github.com/osism/testbed/terraform/inventory.ci\n*deleting   src/github.com/osism/testbed/terraform/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/data.tf\n*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars\n*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/.deploy.ci\n*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci\n",
                            "rc": 0,
                            "stdout_lines": [
                                ".d..t...... src/",
                                ".d..t...... src/github.com/",
                                ".d..t...... src/github.com/osism/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/index",
                                "<f.st...... src/github.com/osism/ansible-collection-commons/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/20/b890adf3f31755b76bb35f6a85b471bc8611a0",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/3f/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/3f/3853c41092b877696fd88dd0b0e64ad9ae191e",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/56/c6294e632c254549c6f8a8f6b2ec82a7ca9475",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/70/87df98cedc09d10112e545be17def164506b0f",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/7e/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/7e/214843694d0dda9b3b5e599d2d1642d7fe94fe",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/b9/f3cb764ba2adef259d06ea75ca42878b99c111",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/e8/f562135696db3b0035ce27bfd6eda9519ec43a",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-commons/.git/objects/fb/f496297b7f2cdde5b294c741e04907868a06f7",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/",
                                "<f.st...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-services/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/index",
                                "<f.st...... src/github.com/osism/ansible-collection-services/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/00/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/03/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/09/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/09/9877719b9c4050e2c4194f238b68435c8d13f9",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/45c18a581110097708fa8077577946117184d5",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1e/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1e/fd16074cc6bea86831de22e4f1ad929170e1ef",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1f/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/1f/7021b98c891b7a22079fc5ce63de4fbb23451a",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/23/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/baa696f26b3d28594fc5f336a58ad086119b51",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/28/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/28/fa13117bed5e1b25592f85da500c9b332ccb8c",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/2a/746d7728a201fefd9e184c48b9ba914e1b6084",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/30/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/30/77a397798a3b682c223ba02a62ac9541c21e80",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/34/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4e/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4e/937145d13b6360a376369f2c4c1b765099004d",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/55/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/72fc5c8da2fbab28c318a47fda85ed94a496d5",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/60/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/61/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/61/6c4bc24d6f449c1e15f37d01d6cfdd30a023cd",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/64/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/65/02fa927ebf5c80a14c6755646873712aa21705",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/68/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/68/52a3a42c25236015b8242dff834c9bc175c9cb",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/6f/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6f/1a7fa6c5b315b8c64d576ec7a74a1fc8b9252f",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/74/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/75/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/75/98a6412096a227921bb9f11312fa1a1aaa44ff",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/7c/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/7c/166b96895d11bfff0947a1710552f256a682fe",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/82/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/44e0d3ddd44b0067ed30791a7f11fdb54dbbdf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/85/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/e8931276538c4be6f59fafe29b8a1ff9f19df0",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/88/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/88/1d5461053b6f86f35c63de6bf89c2601be1b42",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/89/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8a/2ed47a86e336183bf5421866d177b8f7793621",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8b/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/7037cbc10715de30234839daecdb6d338db3a8",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/d796c8699443e29829130f7566d723fdb064c8",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/8d/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/92/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/92/f595b15b5c44505c6375fc780fc8a7581c9f0a",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/95/a1a4c804cb2966bdb38ea0848b0beace188d5c",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/95/eade9661e9f5a2adadcad31f186fd464b69eec",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/9a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/9a/1290e868aa83d5df5a0b03c32e477f221a1680",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a0/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a0/e83416d624fb10a8eae3c94ae66a49b509c14f",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a3/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a4/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a4/68b55c5974bb0598142dcefdc8d03458b56310",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/a5/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a8/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/a8/ef214cc47e25f6f8e3e4e30e3073c767010688",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/aa/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/aa/a04863653a2361deb5b2bc3e6f982d0580281e",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ac/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ac/53de0b381cbe31ed0537e312688c03c3e7317d",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ba/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ba/9a4f0f9c8e0bb623f6a5f00ff5d62bb94a5ec1",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/bd/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/bd/2deeb0500bea24f8cdaf0e918c5f1882045dd2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/c9/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/c9/fa05f95375f6281289163aac2d872e3efb568b",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ce/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/cf/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/cf/fb65c6bb455cd47e135ced1d99a1c28ac93393",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ee/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ee/fe454140e7de3a5ee0dc6c0b6c36322e7ee899",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ef/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ef/7deeb4579605284f54b655ce2adbf14594bd44",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fa/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/fa/47d8a04e45c4b19f035492bdc2e79db626758b",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_yrzn001.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2",
                                "<f.st...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2",
                                ".d..t...... src/github.com/osism/terraform-base/",
                                "<f..t...... src/github.com/osism/terraform-base/.yamllint.yml",
                                "<f..t...... src/github.com/osism/terraform-base/.zuul.yaml",
                                "<f..t...... src/github.com/osism/terraform-base/LICENSE",
                                "<f..t...... src/github.com/osism/terraform-base/README.md",
                                ".d..t...... src/github.com/osism/terraform-base/.git/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config.worktree",
                                "<f..t...... src/github.com/osism/terraform-base/.git/description",
                                "<f..t...... src/github.com/osism/terraform-base/.git/index",
                                "<f..t...... src/github.com/osism/terraform-base/.git/packed-refs",
                                ".d..t...... src/github.com/osism/terraform-base/.git/branches/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/hooks/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/terraform-base/.git/info/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/info/exclude",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/00/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/01/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/02/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/05/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/06/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/07/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/08/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/10/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/12/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/15/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/16/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/17/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/20/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/21/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/24/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/26/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/29/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/36/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/37/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/41/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/43/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/46/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/47/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/48/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/49/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/52/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/55/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/56/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/58/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/61/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/63/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/66/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/67/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/68/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/70/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/72/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/76/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/80/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/81/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/82/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/83/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/84/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/85/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/86/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/88/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/90/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/91/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/92/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/95/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/96/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/98/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/af/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/da/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/db/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/de/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/df/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/info/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.github/",
                                "<f..t...... src/github.com/osism/terraform-base/.github/renovate.json",
                                ".d..t...... src/github.com/osism/terraform-base/playbooks/",
                                "<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/environments/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf",
                                ".d..t...... src/github.com/osism/testbed/",
                                "<f..t...... src/github.com/osism/testbed/.ansible-lint",
                                "<f..t...... src/github.com/osism/testbed/.flake8",
                                "<f..t...... src/github.com/osism/testbed/.gitignore",
                                "<f..t...... src/github.com/osism/testbed/.yamllint.yml",
                                "<f..t...... src/github.com/osism/testbed/.zuul.yaml",
                                "<f..t...... src/github.com/osism/testbed/LICENSE",
                                "<f..t...... src/github.com/osism/testbed/Makefile",
                                "<f..t...... src/github.com/osism/testbed/README.md",
                                "<f..t...... src/github.com/osism/testbed/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/gilt.yml",
                                "<f..t...... src/github.com/osism/testbed/requirements.txt",
                                "<f..t...... src/github.com/osism/testbed/requirements.yml",
                                "<f..t...... src/github.com/osism/testbed/tox.ini",
                                ".d..t...... src/github.com/osism/testbed/.git/",
                                "<f..t...... src/github.com/osism/testbed/.git/HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/config",
                                "<f..t...... src/github.com/osism/testbed/.git/config.worktree",
                                "<f..t...... src/github.com/osism/testbed/.git/description",
                                "<f..t...... src/github.com/osism/testbed/.git/index",
                                "<f..t...... src/github.com/osism/testbed/.git/packed-refs",
                                ".d..t...... src/github.com/osism/testbed/.git/branches/",
                                ".d..t...... src/github.com/osism/testbed/.git/hooks/",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/testbed/.git/info/",
                                "<f..t...... src/github.com/osism/testbed/.git/info/exclude",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/00/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/01/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/02/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/03/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/04/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/05/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/06/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/07/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/08/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/09/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/10/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/11/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/12/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/13/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/14/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/15/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/16/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/17/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/18/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/19/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/20/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/21/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/22/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/23/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/24/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/25/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/26/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/27/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/28/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/29/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/30/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/31/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/32/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/33/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/34/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/35/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/36/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/37/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/38/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/39/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/40/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/41/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/42/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/43/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/44/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/45/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/46/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/47/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/48/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/49/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/50/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/51/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/52/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/53/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/54/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/55/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/56/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/57/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/58/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/59/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/60/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/61/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/64/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/65/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/66/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/67/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/68/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/69/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/70/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/71/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/72/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/73/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/74/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/75/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/76/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/77/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/78/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/79/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/80/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/81/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/82/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/83/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/84/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/85/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/86/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/87/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/90/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/91/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/92/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/93/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/94/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/95/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/96/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/97/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/98/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/af/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/be/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/da/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/db/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/de/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/df/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/info/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/testbed/.git/refs/",
                                ".d..t...... src/github.com/osism/testbed/.github/",
                                "<f..t...... src/github.com/osism/testbed/.github/renovate.json",
                                ".d..t...... src/github.com/osism/testbed/.github/workflows/",
                                "<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml",
                                ".d..t...... src/github.com/osism/testbed/ansible/",
                                "<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/README.md",
                                "<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/files/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/daemons",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf",
                                ".d..t...... src/github.com/osism/testbed/contrib/",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py",
                                "<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql",
                                "<f..t...... src/github.com/osism/testbed/contrib/semver2.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py",
                                ".d..t...... src/github.com/osism/testbed/contrib/ansible/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample",
                                ".d..t...... src/github.com/osism/testbed/contrib/assets/",
                                "<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot",
                                ".d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster",
                                ".d..t...... src/github.com/osism/testbed/contrib/openstack-cli/",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example",
                                ".d..t...... src/github.com/osism/testbed/contrib/ownca/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/",
                                "<f..t...... src/github.com/osism/testbed/environments/.vault_pass",
                                "<f..t...... src/github.com/osism/testbed/environments/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/environments/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph.test/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/templates/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2",
                                ".d..t...... src/github.com/osism/testbed/environments/generic/",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/hosts",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/monitoring/",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/openrc",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst",
                                ".d..t...... src/github.com/osism/testbed/inventory/",
                                "<f..t...... src/github.com/osism/testbed/inventory/10-custom",
                                "<f..t...... src/github.com/osism/testbed/inventory/20-roles",
                                "<f..t...... src/github.com/osism/testbed/inventory/99-overwrite",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/all/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/moduletypes/",
                                "<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/netbox/resources/",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml",
                                ".d..t...... src/github.com/osism/testbed/network/",
                                "<f..t...... src/github.com/osism/testbed/network/iptables.sh",
                                "<f..t...... src/github.com/osism/testbed/network/vxlan.sh",
                                ".d..t...... src/github.com/osism/testbed/playbooks/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/post.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/pre.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml",
                                ".d..t...... src/github.com/osism/testbed/playbooks/templates/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2",
                                ".d..t...... src/github.com/osism/testbed/playbooks/vars/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml",
                                ".d..t...... src/github.com/osism/testbed/scripts/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/include.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/bootstrap/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/check/",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-domain-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/301-openstack-scs-standards.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/customisations/",
                                "<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/deploy/",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/upgrade/",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/terraform/",
                                "<f..t...... src/github.com/osism/testbed/terraform/Makefile",
                                "<f..t...... src/github.com/osism/testbed/terraform/README.md",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/local.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample",
                                ".d..t...... src/github.com/osism/testbed/terraform/environments/",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars",
                                ".d..t...... src/github.com/osism/testbed/terraform/scripts/",
                                "<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/environment",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/",
                                "*deleting   src/github.com/osism/testbed/terraform/variables.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/provider.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/outputs.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nova.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/main.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/local.env",
                                "*deleting   src/github.com/osism/testbed/terraform/inventory.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/data.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars",
                                "*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.deploy.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-6d43-d34b-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:12.862614Z",
                            "start": "2026-03-13T00:00:44.905822Z"
                        },
                        "id": "0242ac1f-6606-6d43-d34b-00000000002d",
                        "name": "Synchronize src repos to workspace directory."
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 9,
            "failures": 0,
            "ignored": 0,
            "ok": 26,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/osism/testbed/playbooks/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T00:02:25.322573Z",
                    "start": "2026-03-13T00:02:14.130080Z"
                },
                "id": "0242ac1f-6606-2de7-578e-000000000002",
                "name": "Pre play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:14.230046Z",
                            "start": "2026-03-13T00:02:14.147916Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:14.279447Z",
                            "start": "2026-03-13T00:02:14.243767Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.423636",
                            "end": "2026-03-13 00:02:16.675000",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-2de7-578e-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:02:15.251364",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-13 00:02:15 - clean up servers\n2026-03-13 00:02:16 - clean up keypairs\n2026-03-13 00:02:16 - wait for servers to be gone\n2026-03-13 00:02:16 - clean up ports\n2026-03-13 00:02:16 - clean up volumes\n2026-03-13 00:02:16 - disconnect routers\n2026-03-13 00:02:16 - clean up subnets\n2026-03-13 00:02:16 - clean up networks\n2026-03-13 00:02:16 - clean up security groups\n2026-03-13 00:02:16 - clean up floating ips\n2026-03-13 00:02:16 - clean up routers",
                            "stdout_lines": [
                                "2026-03-13 00:02:15 - clean up servers",
                                "2026-03-13 00:02:16 - clean up keypairs",
                                "2026-03-13 00:02:16 - wait for servers to be gone",
                                "2026-03-13 00:02:16 - clean up ports",
                                "2026-03-13 00:02:16 - clean up volumes",
                                "2026-03-13 00:02:16 - disconnect routers",
                                "2026-03-13 00:02:16 - clean up subnets",
                                "2026-03-13 00:02:16 - clean up networks",
                                "2026-03-13 00:02:16 - clean up security groups",
                                "2026-03-13 00:02:16 - clean up floating ips",
                                "2026-03-13 00:02:16 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-2de7-578e-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:16.867839Z",
                            "start": "2026-03-13T00:02:14.350149Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000006",
                        "name": "Clean the cloud environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.unarchive",
                            "changed": false,
                            "dest": "/home/zuul-testbed03",
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "handler": "ZipArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": null,
                                    "decrypt": true,
                                    "dest": "/home/zuul-testbed03",
                                    "exclude": [],
                                    "extra_opts": [],
                                    "group": null,
                                    "include": [
                                        "tofu"
                                    ],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/opentofu/opentofu/releases/download/v1.11.5/tofu_1.11.5_linux_amd64.zip",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed03",
                            "size": 4096,
                            "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1773360136.9365335-13-133415532853516/tofu_1.11.5_linux_amd64qtri7x0o.zip",
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:23.921097Z",
                            "start": "2026-03-13T00:02:16.883902Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000007",
                        "name": "Extract tofu binary"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "rsync",
                                "-avz",
                                "/home/zuul-testbed03/src/github.com/osism/terraform-base/testbed-default/",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform"
                            ],
                            "delta": "0:00:00.058227",
                            "end": "2026-03-13 00:02:24.294004",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "rsync -avz /home/zuul-testbed03/src/github.com/osism/terraform-base/testbed-default/ /home/zuul-testbed03/src/github.com/osism/testbed/terraform",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-2de7-578e-000000000008-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:02:24.235777",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sending incremental file list\ndata.tf\nmain.tf\nmanager.tf\nneutron.tf\nnodes.tf\nnova.tf\noutputs.tf\nprovider.tf\nvariables.tf\ncustomisations/\ncustomisations/access_floatingip_custom.tf\ncustomisations/access_ipv4_custom.tf\ncustomisations/access_ipv6_custom.tf\ncustomisations/default_custom.tf\ncustomisations/external_api_custom.tf\ncustomisations/neutron_floatingip_custom.tf\noverrides/\noverrides/manager_boot_from_image_override.tf\noverrides/manager_boot_from_volume_override.tf\noverrides/neutron_availability_zone_hints_network_override.tf\noverrides/neutron_availability_zone_hints_router_override.tf\noverrides/neutron_router_enable_snat_override.tf\noverrides/nodes_boot_from_image_override.tf\noverrides/nodes_boot_from_volume_override.tf\noverrides/nodes_use_ephemeral_storage_override.tf\n\nsent 6,928 bytes  received 470 bytes  14,796.00 bytes/sec\ntotal size is 26,532  speedup is 3.59",
                            "stdout_lines": [
                                "sending incremental file list",
                                "data.tf",
                                "main.tf",
                                "manager.tf",
                                "neutron.tf",
                                "nodes.tf",
                                "nova.tf",
                                "outputs.tf",
                                "provider.tf",
                                "variables.tf",
                                "customisations/",
                                "customisations/access_floatingip_custom.tf",
                                "customisations/access_ipv4_custom.tf",
                                "customisations/access_ipv6_custom.tf",
                                "customisations/default_custom.tf",
                                "customisations/external_api_custom.tf",
                                "customisations/neutron_floatingip_custom.tf",
                                "overrides/",
                                "overrides/manager_boot_from_image_override.tf",
                                "overrides/manager_boot_from_volume_override.tf",
                                "overrides/neutron_availability_zone_hints_network_override.tf",
                                "overrides/neutron_availability_zone_hints_router_override.tf",
                                "overrides/neutron_router_enable_snat_override.tf",
                                "overrides/nodes_boot_from_image_override.tf",
                                "overrides/nodes_boot_from_volume_override.tf",
                                "overrides/nodes_use_ephemeral_storage_override.tf",
                                "",
                                "sent 6,928 bytes  received 470 bytes  14,796.00 bytes/sec",
                                "total size is 26,532  speedup is 3.59"
                            ],
                            "zuul_log_id": "0242ac1f-6606-2de7-578e-000000000008-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:24.494209Z",
                            "start": "2026-03-13T00:02:23.967718Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000008",
                        "name": "Sync terraform blueprint"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.template",
                            "changed": true,
                            "checksum": "c7d87fff32e4ed8365964dfc6d78aaf197c62596",
                            "dest": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/local.env",
                            "diff": [],
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "local.env.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "c7d87fff32e4ed8365964dfc6d78aaf197c62596",
                                    "content": null,
                                    "dest": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/local.env",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": "0644",
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1773360144.5578966-22-51137276045201/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "8159794407094435d0565858fa369f4a",
                            "mode": "0644",
                            "owner": "zuul-testbed03",
                            "size": 42,
                            "src": "/home/zuul-testbed03/.ansible/tmp/ansible-tmp-1773360144.5578966-22-51137276045201/source",
                            "state": "file",
                            "uid": 1003
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:25.322573Z",
                            "start": "2026-03-13T00:02:24.503151Z"
                        },
                        "id": "0242ac1f-6606-2de7-578e-000000000009",
                        "name": "Create local.env file"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 3,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/osism/testbed/playbooks/deploy.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:38:44.001374Z",
                    "start": "2026-03-13T00:02:26.305979Z"
                },
                "id": "0242ac1f-6606-6cb6-5436-000000000002",
                "name": "Deploy testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:26.404218Z",
                            "start": "2026-03-13T00:02:26.321967Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000004",
                        "name": "Print used ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version": "2025.1",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:26.476520Z",
                            "start": "2026-03-13T00:02:26.413629Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000005",
                        "name": "Print used openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version": "latest",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:26.544056Z",
                            "start": "2026-03-13T00:02:26.483452Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000006",
                        "name": "Print used manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "image_username": "ubuntu"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:26.643563Z",
                            "start": "2026-03-13T00:02:26.551272Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000007",
                        "name": "Set facts (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:02:26.688048Z",
                            "start": "2026-03-13T00:02:26.652236Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000008",
                        "name": "Set facts (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "make",
                                "ENVIRONMENT=ci-ubuntu-24.04",
                                "CLOUD=ci",
                                "IMAGE_USERNAME=ubuntu",
                                "TERRAFORM=/home/zuul-testbed03/tofu",
                                "VERSION_CEPH=reef",
                                "VERSION_MANAGER=latest",
                                "VERSION_OPENSTACK=2025.1",
                                "TEMPEST=true",
                                "CEPH_STACK=ceph-ansible",
                                "create"
                            ],
                            "delta": "0:01:39.378750",
                            "end": "2026-03-13 00:04:06.951164",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "make\n  ENVIRONMENT=ci-ubuntu-24.04\n  CLOUD=ci\n  IMAGE_USERNAME=ubuntu\n  TERRAFORM=/home/zuul-testbed03/tofu\n  VERSION_CEPH=reef\n  VERSION_MANAGER=latest\n  VERSION_OPENSTACK=2025.1\n  TEMPEST=true\n  CEPH_STACK=ceph-ansible\n  create\n",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:02:27.572414",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\nInitializing the backend...\n\nInitializing provider plugins...\n- terraform.io/builtin/terraform is built in to OpenTofu\n- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...\n- Finding hashicorp/local versions matching \">= 2.2.0\"...\n- Finding latest version of hashicorp/null...\n- Installing terraform-provider-openstack/openstack v3.4.0...\n- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)\n- Installing hashicorp/local v2.7.0...\n- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)\n- Installing hashicorp/null v3.2.4...\n- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)\n\nProviders are signed by their developers.\nIf you'd like to know more about provider signing, you can read about it here:\nhttps://opentofu.org/docs/cli/plugins/signing/\n\nOpenTofu has created a lock file .terraform.lock.hcl to record the provider\nselections it made above. Include this file in your version control repository\nso that OpenTofu can guarantee to make the same selections by default when\nyou run \"tofu init\" in the future.\n\nOpenTofu has been successfully initialized!\n\nYou may now begin working with OpenTofu. Try running \"tofu plan\" to see\nany changes that are required for your infrastructure. All OpenTofu commands\nshould now work.\n\nIf you ever set or change modules or backend configuration for OpenTofu,\nrerun this command to reinitialize your working directory. If you forget, other\ncommands will detect it and remind you to do so if necessary.\nCreated and switched to workspace \"ci\"!\n\nYou're now on a new, empty workspace. Workspaces isolate their state,\nso if you run \"tofu plan\" OpenTofu will not see any existing state\nfor this configuration.\nci.auto.tfvars\ndefault_custom.tf\ndata.openstack_networking_network_v2.public: Reading...\ndata.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]\n\nOpenTofu used the selected providers to generate the following execution\nplan. Resource actions are indicated with the following symbols:\n  + create\n <= read (data resources)\n\nOpenTofu will perform the following actions:\n\n  # data.openstack_images_image_v2.image will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # data.openstack_images_image_v2.image_node will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image_node\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # local_file.MANAGER_ADDRESS will be created\n  + resource \"local_file\" \"MANAGER_ADDRESS\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".MANAGER_ADDRESS.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.id_rsa_pub will be created\n  + resource \"local_file\" \"id_rsa_pub\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".id_rsa.ci.pub\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.inventory will be created\n  + resource \"local_file\" \"inventory\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \"inventory.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_sensitive_file.id_rsa will be created\n  + resource \"local_sensitive_file\" \"id_rsa\" {\n      + content              = (sensitive value)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0700\"\n      + file_permission      = \"0600\"\n      + filename             = \".id_rsa.ci\"\n      + id                   = (known after apply)\n    }\n\n  # null_resource.node_semaphore will be created\n  + resource \"null_resource\" \"node_semaphore\" {\n      + id = (known after apply)\n    }\n\n  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-manager-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[6] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-6-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[7] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-7-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[8] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-8-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_compute_instance_v2.manager_server will be created\n  + resource \"openstack_compute_instance_v2\" \"manager_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-4V-16\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-manager\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = (sensitive value)\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[0] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-0\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[1] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-1\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[2] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-2\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[3] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-3\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[4] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-4\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[5] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-5\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_keypair_v2.key will be created\n  + resource \"openstack_compute_keypair_v2\" \"key\" {\n      + fingerprint = (known after apply)\n      + id          = (known after apply)\n      + name        = \"testbed\"\n      + private_key = (sensitive value)\n      + public_key  = (known after apply)\n      + region      = (known after apply)\n      + user_id     = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created\n  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {\n      + fixed_ip    = (known after apply)\n      + floating_ip = (known after apply)\n      + id          = (known after apply)\n      + port_id     = (known after apply)\n      + region      = (known after apply)\n    }\n\n  # openstack_networking_floatingip_v2.manager_floating_ip will be created\n  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {\n      + address    = (known after apply)\n      + all_tags   = (known after apply)\n      + dns_domain = (known after apply)\n      + dns_name   = (known after apply)\n      + fixed_ip   = (known after apply)\n      + id         = (known after apply)\n      + pool       = \"public\"\n      + port_id    = (known after apply)\n      + region     = (known after apply)\n      + subnet_id  = (known after apply)\n      + tenant_id  = (known after apply)\n    }\n\n  # openstack_networking_network_v2.net_management will be created\n  + resource \"openstack_networking_network_v2\" \"net_management\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + dns_domain              = (known after apply)\n      + external                = (known after apply)\n      + id                      = (known after apply)\n      + mtu                     = (known after apply)\n      + name                    = \"net-testbed-management\"\n      + port_security_enabled   = (known after apply)\n      + qos_policy_id           = (known after apply)\n      + region                  = (known after apply)\n      + shared                  = (known after apply)\n      + tenant_id               = (known after apply)\n      + transparent_vlan        = (known after apply)\n\n      + segments (known after apply)\n    }\n\n  # openstack_networking_port_v2.manager_port_management will be created\n  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.5\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[0] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.10\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[1] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.11\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[2] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.12\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[3] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.13\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[4] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.14\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[5] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.15\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_router_interface_v2.router_interface will be created\n  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {\n      + force_destroy = false\n      + id            = (known after apply)\n      + port_id       = (known after apply)\n      + region        = (known after apply)\n      + router_id     = (known after apply)\n      + subnet_id     = (known after apply)\n    }\n\n  # openstack_networking_router_v2.router will be created\n  + resource \"openstack_networking_router_v2\" \"router\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + distributed             = (known after apply)\n      + enable_snat             = (known after apply)\n      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"\n      + external_qos_policy_id  = (known after apply)\n      + id                      = (known after apply)\n      + name                    = \"testbed\"\n      + region                  = (known after apply)\n      + tenant_id               = (known after apply)\n\n      + external_fixed_ip (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {\n      + description             = \"ssh\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 22\n      + port_range_min          = 22\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {\n      + description             = \"wireguard\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 51820\n      + port_range_min          = 51820\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {\n      + description             = \"vrrp\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"112\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_management will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {\n      + all_tags    = (known after apply)\n      + description = \"management security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-management\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_node will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {\n      + all_tags    = (known after apply)\n      + description = \"node security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-node\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_subnet_v2.subnet_management will be created\n  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {\n      + all_tags          = (known after apply)\n      + cidr              = \"192.168.16.0/20\"\n      + dns_nameservers   = [\n          + \"8.8.8.8\",\n          + \"9.9.9.9\",\n        ]\n      + enable_dhcp       = true\n      + gateway_ip        = (known after apply)\n      + id                = (known after apply)\n      + ip_version        = 4\n      + ipv6_address_mode = (known after apply)\n      + ipv6_ra_mode      = (known after apply)\n      + name              = \"subnet-testbed-management\"\n      + network_id        = (known after apply)\n      + no_gateway        = false\n      + region            = (known after apply)\n      + service_types     = (known after apply)\n      + tenant_id         = (known after apply)\n\n      + allocation_pool {\n          + end   = \"192.168.31.250\"\n          + start = \"192.168.31.200\"\n        }\n    }\n\n  # terraform_data.image will be created\n  + resource \"terraform_data\" \"image\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\n  # terraform_data.image_node will be created\n  + resource \"terraform_data\" \"image_node\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\nPlan: 64 to add, 0 to change, 0 to destroy.\n\nChanges to Outputs:\n  + manager_address = (sensitive value)\n  + private_key     = (sensitive value)\nterraform_data.image: Creating...\nterraform_data.image: Creation complete after 0s [id=78bb27fa-0e4d-9715-f8e0-6161647c58c6]\nterraform_data.image_node: Creating...\nterraform_data.image_node: Creation complete after 0s [id=44ac8a12-1240-12da-59b0-177ac206e54b]\ndata.openstack_images_image_v2.image: Reading...\ndata.openstack_images_image_v2.image_node: Reading...\nopenstack_compute_keypair_v2.key: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creating...\nopenstack_networking_network_v2.net_management: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creating...\ndata.openstack_images_image_v2.image: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creating...\ndata.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[4]: Creating...\nopenstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creating...\nopenstack_networking_network_v2.net_management: Creation complete after 1s [id=51b7d847-e6cc-4f71-a38f-118bdb255db6]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=9392309c-12b1-408e-a498-5a5dc61ed8cc]\nopenstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=f4a99b06-bc15-4835-b50d-6358b82edd6b]\nopenstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=da000f05-04bb-42ea-839f-c75cda06f5f4]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=77efb09a-3c5f-4577-ba7d-fc82392d2604]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=f4e98a13-908f-47c9-b3df-604b890733f1]\nopenstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=9de382f4-b65a-458b-8cc5-b0eb7440fe81]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creating...\nlocal_file.id_rsa_pub: Creating...\nlocal_file.id_rsa_pub: Creation complete after 0s [id=3e07636b3084577c6c907acc34eae13efa6f7761]\nlocal_sensitive_file.id_rsa: Creating...\nlocal_sensitive_file.id_rsa: Creation complete after 0s [id=00332d9d63cc805a7b8a58ff92558bbcca8be785]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=45a6e807-92ae-410a-8871-af67928fbfa9]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creating...\nopenstack_networking_subnet_v2.subnet_management: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=de104e6f-6514-4ab5-a08d-43193a3bf917]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=7a80a5ff-f3bf-405b-902d-2c7a644821a3]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=96c01f28-dfe3-49e3-bcac-8f6c78683f93]\nopenstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=e279997b-a9ed-4720-a660-f6a14b5bdd39]\nopenstack_networking_router_v2.router: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=dd0e771c-a216-451f-be9a-7a9338004557]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=f673daef-4b26-4c19-9ad2-6e63015dc391]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=b0e3959e-31e2-4ee1-98d8-465e3bcb125b]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=f3b73d97-a549-411a-96e1-9b740e8b4b02]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=bb461c46-8762-4442-a15b-8a2831c93e44]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=180093e4-fb06-417b-b075-79800beac41f]\nopenstack_networking_router_v2.router: Creation complete after 4s [id=afaa0451-156b-4ec6-9c75-47f62b5f7fc5]\nopenstack_networking_secgroup_v2.security_group_management: Creating...\nopenstack_networking_router_interface_v2.router_interface: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creation complete after 1s [id=afe10a17-39a4-4e9c-90db-edab85aac2cc]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...\nopenstack_networking_port_v2.node_port_management[3]: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...\nopenstack_networking_port_v2.node_port_management[4]: Creating...\nopenstack_networking_port_v2.node_port_management[0]: Creating...\nopenstack_networking_port_v2.node_port_management[5]: Creating...\nopenstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=019211b1-712d-417d-9bfe-c1c08f7461c8]\nopenstack_networking_port_v2.node_port_management[2]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 0s [id=9e647355-a6bb-43d2-b482-6d9681eae497]\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=a3fd03c2-f311-413e-87dc-4b3629893332]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...\nopenstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=0e20cc2d-f5c4-43e4-bf07-70cac2a883f1]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...\nopenstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=cd46a55b-3127-464a-845c-fe24331a784d]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...\nopenstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=91da1218-c10d-476c-bd3a-07c5008fc822]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=7c9a6544-5e48-46fa-99db-f27f04c7d5c8]\nopenstack_networking_port_v2.manager_port_management: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=9d8018a8-bfca-489a-9cb1-829482b39b1a]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...\nopenstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=38bd7b83-ef2a-4fd5-b9d7-0ddfaf83a6de]\nopenstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=b05cadca-a1ba-40c8-8416-3d17d0ec1cf4]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 0s [id=a8e8c225-e3e7-42bb-b36c-e512b925b3b5]\nopenstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=c7ea0029-231a-461d-987b-fe01ea7248e2]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 2s [id=3c2ac8be-08f8-48d9-90f6-156e13e48375]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=0f2464ed-24d8-47c3-bf5d-353c7478c115]\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 2s [id=b239b0d1-4ab5-4b83-bc9f-49502f734017]\nopenstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=670071ab-2c27-48c5-a098-e705a52445bb]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=ca2632ca-4020-4095-8a16-bc6e6949e9a8]\nopenstack_networking_router_interface_v2.router_interface: Creation complete after 4s [id=a3eed546-ab51-45a7-9e06-1f6b54c5c1e0]\nopenstack_networking_floatingip_v2.manager_floating_ip: Creating...\nopenstack_compute_instance_v2.node_server[2]: Creating...\nopenstack_compute_instance_v2.node_server[1]: Creating...\nopenstack_compute_instance_v2.node_server[3]: Creating...\nopenstack_compute_instance_v2.node_server[4]: Creating...\nopenstack_compute_instance_v2.node_server[0]: Creating...\nopenstack_compute_instance_v2.node_server[5]: Creating...\nopenstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=46349f6f-b013-416a-92af-fc74b2d1d16d]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...\nlocal_file.MANAGER_ADDRESS: Creating...\nlocal_file.inventory: Creating...\nlocal_file.inventory: Creation complete after 0s [id=7fdf76c861c253bb1991f817d102f69295129a60]\nlocal_file.MANAGER_ADDRESS: Creation complete after 0s [id=a135100d5da4d623c3f27077fdd081752abf20d8]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=46349f6f-b013-416a-92af-fc74b2d1d16d]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=8fb45040-5735-41a2-8689-07a3f0b92e71]\nopenstack_compute_instance_v2.node_server[4]: Creation complete after 31s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [40s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Creation complete after 42s [id=a7573d11-6cf5-4154-9a46-d61d5a6dcb3e]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [50s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Creation complete after 51s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a]\nopenstack_compute_instance_v2.node_server[5]: Creation complete after 51s [id=9c087d67-da7c-4387-b854-0f22f069bd28]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [1m0s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Creation complete after 1m1s [id=4c1d0bbb-9cef-4ff3-8bfb-8f24cbacbb49]\nnull_resource.node_semaphore: Creating...\nnull_resource.node_semaphore: Creation complete after 0s [id=5916097257062349066]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...\nopenstack_compute_instance_v2.manager_server: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 4s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/45a6e807-92ae-410a-8871-af67928fbfa9]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 3s [id=9c087d67-da7c-4387-b854-0f22f069bd28/da000f05-04bb-42ea-839f-c75cda06f5f4]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 4s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/de104e6f-6514-4ab5-a08d-43193a3bf917]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 4s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/77efb09a-3c5f-4577-ba7d-fc82392d2604]\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 9s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/f4a99b06-bc15-4835-b50d-6358b82edd6b]\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 9s [id=9c087d67-da7c-4387-b854-0f22f069bd28/9392309c-12b1-408e-a498-5a5dc61ed8cc]\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/f4e98a13-908f-47c9-b3df-604b890733f1]\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=9c087d67-da7c-4387-b854-0f22f069bd28/9de382f4-b65a-458b-8cc5-b0eb7440fe81]\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 9s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/7a80a5ff-f3bf-405b-902d-2c7a644821a3]\nopenstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.manager_server: Creation complete after 20s [id=3e8dbe75-a73d-4282-b61b-1bdcc2b3c46d]\n\nApply complete! Resources: 64 added, 0 changed, 0 destroyed.\n\nOutputs:\n\nmanager_address = <sensitive>\nprivate_key = <sensitive>",
                            "stdout_lines": [
                                "",
                                "Initializing the backend...",
                                "",
                                "Initializing provider plugins...",
                                "- terraform.io/builtin/terraform is built in to OpenTofu",
                                "- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...",
                                "- Finding hashicorp/local versions matching \">= 2.2.0\"...",
                                "- Finding latest version of hashicorp/null...",
                                "- Installing terraform-provider-openstack/openstack v3.4.0...",
                                "- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)",
                                "- Installing hashicorp/local v2.7.0...",
                                "- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing hashicorp/null v3.2.4...",
                                "- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)",
                                "",
                                "Providers are signed by their developers.",
                                "If you'd like to know more about provider signing, you can read about it here:",
                                "https://opentofu.org/docs/cli/plugins/signing/",
                                "",
                                "OpenTofu has created a lock file .terraform.lock.hcl to record the provider",
                                "selections it made above. Include this file in your version control repository",
                                "so that OpenTofu can guarantee to make the same selections by default when",
                                "you run \"tofu init\" in the future.",
                                "",
                                "OpenTofu has been successfully initialized!",
                                "",
                                "You may now begin working with OpenTofu. Try running \"tofu plan\" to see",
                                "any changes that are required for your infrastructure. All OpenTofu commands",
                                "should now work.",
                                "",
                                "If you ever set or change modules or backend configuration for OpenTofu,",
                                "rerun this command to reinitialize your working directory. If you forget, other",
                                "commands will detect it and remind you to do so if necessary.",
                                "Created and switched to workspace \"ci\"!",
                                "",
                                "You're now on a new, empty workspace. Workspaces isolate their state,",
                                "so if you run \"tofu plan\" OpenTofu will not see any existing state",
                                "for this configuration.",
                                "ci.auto.tfvars",
                                "default_custom.tf",
                                "data.openstack_networking_network_v2.public: Reading...",
                                "data.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]",
                                "",
                                "OpenTofu used the selected providers to generate the following execution",
                                "plan. Resource actions are indicated with the following symbols:",
                                "  + create",
                                " <= read (data resources)",
                                "",
                                "OpenTofu will perform the following actions:",
                                "",
                                "  # data.openstack_images_image_v2.image will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # data.openstack_images_image_v2.image_node will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image_node\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.MANAGER_ADDRESS will be created",
                                "  + resource \"local_file\" \"MANAGER_ADDRESS\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".MANAGER_ADDRESS.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.id_rsa_pub will be created",
                                "  + resource \"local_file\" \"id_rsa_pub\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".id_rsa.ci.pub\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.inventory will be created",
                                "  + resource \"local_file\" \"inventory\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \"inventory.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_sensitive_file.id_rsa will be created",
                                "  + resource \"local_sensitive_file\" \"id_rsa\" {",
                                "      + content              = (sensitive value)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0700\"",
                                "      + file_permission      = \"0600\"",
                                "      + filename             = \".id_rsa.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # null_resource.node_semaphore will be created",
                                "  + resource \"null_resource\" \"node_semaphore\" {",
                                "      + id = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-manager-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[6] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-6-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[7] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-7-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[8] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-8-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.manager_server will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"manager_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-4V-16\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-manager\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = (sensitive value)",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[0] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-0\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[1] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-1\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[2] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-2\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[3] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-3\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[4] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-4\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[5] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-5\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_keypair_v2.key will be created",
                                "  + resource \"openstack_compute_keypair_v2\" \"key\" {",
                                "      + fingerprint = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed\"",
                                "      + private_key = (sensitive value)",
                                "      + public_key  = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + user_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created",
                                "  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {",
                                "      + fixed_ip    = (known after apply)",
                                "      + floating_ip = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + port_id     = (known after apply)",
                                "      + region      = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_v2.manager_floating_ip will be created",
                                "  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {",
                                "      + address    = (known after apply)",
                                "      + all_tags   = (known after apply)",
                                "      + dns_domain = (known after apply)",
                                "      + dns_name   = (known after apply)",
                                "      + fixed_ip   = (known after apply)",
                                "      + id         = (known after apply)",
                                "      + pool       = \"public\"",
                                "      + port_id    = (known after apply)",
                                "      + region     = (known after apply)",
                                "      + subnet_id  = (known after apply)",
                                "      + tenant_id  = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_network_v2.net_management will be created",
                                "  + resource \"openstack_networking_network_v2\" \"net_management\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + dns_domain              = (known after apply)",
                                "      + external                = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + mtu                     = (known after apply)",
                                "      + name                    = \"net-testbed-management\"",
                                "      + port_security_enabled   = (known after apply)",
                                "      + qos_policy_id           = (known after apply)",
                                "      + region                  = (known after apply)",
                                "      + shared                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "      + transparent_vlan        = (known after apply)",
                                "",
                                "      + segments (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.manager_port_management will be created",
                                "  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.5\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[0] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.10\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[1] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.11\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[2] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.12\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[3] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.13\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[4] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.14\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[5] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.15\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_router_interface_v2.router_interface will be created",
                                "  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {",
                                "      + force_destroy = false",
                                "      + id            = (known after apply)",
                                "      + port_id       = (known after apply)",
                                "      + region        = (known after apply)",
                                "      + router_id     = (known after apply)",
                                "      + subnet_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_router_v2.router will be created",
                                "  + resource \"openstack_networking_router_v2\" \"router\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + distributed             = (known after apply)",
                                "      + enable_snat             = (known after apply)",
                                "      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"",
                                "      + external_qos_policy_id  = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + name                    = \"testbed\"",
                                "      + region                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "",
                                "      + external_fixed_ip (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {",
                                "      + description             = \"ssh\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 22",
                                "      + port_range_min          = 22",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {",
                                "      + description             = \"wireguard\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 51820",
                                "      + port_range_min          = 51820",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {",
                                "      + description             = \"vrrp\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"112\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_management will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"management security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-management\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_node will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"node security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-node\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_subnet_v2.subnet_management will be created",
                                "  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {",
                                "      + all_tags          = (known after apply)",
                                "      + cidr              = \"192.168.16.0/20\"",
                                "      + dns_nameservers   = [",
                                "          + \"8.8.8.8\",",
                                "          + \"9.9.9.9\",",
                                "        ]",
                                "      + enable_dhcp       = true",
                                "      + gateway_ip        = (known after apply)",
                                "      + id                = (known after apply)",
                                "      + ip_version        = 4",
                                "      + ipv6_address_mode = (known after apply)",
                                "      + ipv6_ra_mode      = (known after apply)",
                                "      + name              = \"subnet-testbed-management\"",
                                "      + network_id        = (known after apply)",
                                "      + no_gateway        = false",
                                "      + region            = (known after apply)",
                                "      + service_types     = (known after apply)",
                                "      + tenant_id         = (known after apply)",
                                "",
                                "      + allocation_pool {",
                                "          + end   = \"192.168.31.250\"",
                                "          + start = \"192.168.31.200\"",
                                "        }",
                                "    }",
                                "",
                                "  # terraform_data.image will be created",
                                "  + resource \"terraform_data\" \"image\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "  # terraform_data.image_node will be created",
                                "  + resource \"terraform_data\" \"image_node\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "Plan: 64 to add, 0 to change, 0 to destroy.",
                                "",
                                "Changes to Outputs:",
                                "  + manager_address = (sensitive value)",
                                "  + private_key     = (sensitive value)",
                                "terraform_data.image: Creating...",
                                "terraform_data.image: Creation complete after 0s [id=78bb27fa-0e4d-9715-f8e0-6161647c58c6]",
                                "terraform_data.image_node: Creating...",
                                "terraform_data.image_node: Creation complete after 0s [id=44ac8a12-1240-12da-59b0-177ac206e54b]",
                                "data.openstack_images_image_v2.image: Reading...",
                                "data.openstack_images_image_v2.image_node: Reading...",
                                "openstack_compute_keypair_v2.key: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creating...",
                                "openstack_networking_network_v2.net_management: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creating...",
                                "data.openstack_images_image_v2.image: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creating...",
                                "data.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creating...",
                                "openstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creating...",
                                "openstack_networking_network_v2.net_management: Creation complete after 1s [id=51b7d847-e6cc-4f71-a38f-118bdb255db6]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 4s [id=9392309c-12b1-408e-a498-5a5dc61ed8cc]",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=f4a99b06-bc15-4835-b50d-6358b82edd6b]",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=da000f05-04bb-42ea-839f-c75cda06f5f4]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=77efb09a-3c5f-4577-ba7d-fc82392d2604]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=f4e98a13-908f-47c9-b3df-604b890733f1]",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=9de382f4-b65a-458b-8cc5-b0eb7440fe81]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creating...",
                                "local_file.id_rsa_pub: Creating...",
                                "local_file.id_rsa_pub: Creation complete after 0s [id=3e07636b3084577c6c907acc34eae13efa6f7761]",
                                "local_sensitive_file.id_rsa: Creating...",
                                "local_sensitive_file.id_rsa: Creation complete after 0s [id=00332d9d63cc805a7b8a58ff92558bbcca8be785]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=45a6e807-92ae-410a-8871-af67928fbfa9]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creating...",
                                "openstack_networking_subnet_v2.subnet_management: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 3s [id=de104e6f-6514-4ab5-a08d-43193a3bf917]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=7a80a5ff-f3bf-405b-902d-2c7a644821a3]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=96c01f28-dfe3-49e3-bcac-8f6c78683f93]",
                                "openstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=e279997b-a9ed-4720-a660-f6a14b5bdd39]",
                                "openstack_networking_router_v2.router: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=dd0e771c-a216-451f-be9a-7a9338004557]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 3s [id=f673daef-4b26-4c19-9ad2-6e63015dc391]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=b0e3959e-31e2-4ee1-98d8-465e3bcb125b]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=f3b73d97-a549-411a-96e1-9b740e8b4b02]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 3s [id=bb461c46-8762-4442-a15b-8a2831c93e44]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 3s [id=180093e4-fb06-417b-b075-79800beac41f]",
                                "openstack_networking_router_v2.router: Creation complete after 4s [id=afaa0451-156b-4ec6-9c75-47f62b5f7fc5]",
                                "openstack_networking_secgroup_v2.security_group_management: Creating...",
                                "openstack_networking_router_interface_v2.router_interface: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creation complete after 1s [id=afe10a17-39a4-4e9c-90db-edab85aac2cc]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...",
                                "openstack_networking_port_v2.node_port_management[3]: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...",
                                "openstack_networking_port_v2.node_port_management[4]: Creating...",
                                "openstack_networking_port_v2.node_port_management[0]: Creating...",
                                "openstack_networking_port_v2.node_port_management[5]: Creating...",
                                "openstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=019211b1-712d-417d-9bfe-c1c08f7461c8]",
                                "openstack_networking_port_v2.node_port_management[2]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 0s [id=9e647355-a6bb-43d2-b482-6d9681eae497]",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=a3fd03c2-f311-413e-87dc-4b3629893332]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...",
                                "openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=0e20cc2d-f5c4-43e4-bf07-70cac2a883f1]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...",
                                "openstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=cd46a55b-3127-464a-845c-fe24331a784d]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...",
                                "openstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=91da1218-c10d-476c-bd3a-07c5008fc822]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=7c9a6544-5e48-46fa-99db-f27f04c7d5c8]",
                                "openstack_networking_port_v2.manager_port_management: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=9d8018a8-bfca-489a-9cb1-829482b39b1a]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...",
                                "openstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=38bd7b83-ef2a-4fd5-b9d7-0ddfaf83a6de]",
                                "openstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=b05cadca-a1ba-40c8-8416-3d17d0ec1cf4]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 0s [id=a8e8c225-e3e7-42bb-b36c-e512b925b3b5]",
                                "openstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=c7ea0029-231a-461d-987b-fe01ea7248e2]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 2s [id=3c2ac8be-08f8-48d9-90f6-156e13e48375]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=0f2464ed-24d8-47c3-bf5d-353c7478c115]",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 2s [id=b239b0d1-4ab5-4b83-bc9f-49502f734017]",
                                "openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=670071ab-2c27-48c5-a098-e705a52445bb]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=ca2632ca-4020-4095-8a16-bc6e6949e9a8]",
                                "openstack_networking_router_interface_v2.router_interface: Creation complete after 4s [id=a3eed546-ab51-45a7-9e06-1f6b54c5c1e0]",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creating...",
                                "openstack_compute_instance_v2.node_server[2]: Creating...",
                                "openstack_compute_instance_v2.node_server[1]: Creating...",
                                "openstack_compute_instance_v2.node_server[3]: Creating...",
                                "openstack_compute_instance_v2.node_server[4]: Creating...",
                                "openstack_compute_instance_v2.node_server[0]: Creating...",
                                "openstack_compute_instance_v2.node_server[5]: Creating...",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 2s [id=46349f6f-b013-416a-92af-fc74b2d1d16d]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...",
                                "local_file.MANAGER_ADDRESS: Creating...",
                                "local_file.inventory: Creating...",
                                "local_file.inventory: Creation complete after 0s [id=7fdf76c861c253bb1991f817d102f69295129a60]",
                                "local_file.MANAGER_ADDRESS: Creation complete after 0s [id=a135100d5da4d623c3f27077fdd081752abf20d8]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 2s [id=46349f6f-b013-416a-92af-fc74b2d1d16d]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=8fb45040-5735-41a2-8689-07a3f0b92e71]",
                                "openstack_compute_instance_v2.node_server[4]: Creation complete after 31s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [40s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Creation complete after 42s [id=a7573d11-6cf5-4154-9a46-d61d5a6dcb3e]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [50s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Creation complete after 51s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a]",
                                "openstack_compute_instance_v2.node_server[5]: Creation complete after 51s [id=9c087d67-da7c-4387-b854-0f22f069bd28]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [1m0s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Creation complete after 1m1s [id=4c1d0bbb-9cef-4ff3-8bfb-8f24cbacbb49]",
                                "null_resource.node_semaphore: Creating...",
                                "null_resource.node_semaphore: Creation complete after 0s [id=5916097257062349066]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...",
                                "openstack_compute_instance_v2.manager_server: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 4s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/45a6e807-92ae-410a-8871-af67928fbfa9]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 3s [id=9c087d67-da7c-4387-b854-0f22f069bd28/da000f05-04bb-42ea-839f-c75cda06f5f4]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 4s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/de104e6f-6514-4ab5-a08d-43193a3bf917]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 4s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/77efb09a-3c5f-4577-ba7d-fc82392d2604]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 9s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/f4a99b06-bc15-4835-b50d-6358b82edd6b]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 9s [id=9c087d67-da7c-4387-b854-0f22f069bd28/9392309c-12b1-408e-a498-5a5dc61ed8cc]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=d23d647f-999e-4b3a-b05b-0c06200f4e9a/f4e98a13-908f-47c9-b3df-604b890733f1]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=9c087d67-da7c-4387-b854-0f22f069bd28/9de382f4-b65a-458b-8cc5-b0eb7440fe81]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 9s [id=f72ba3f3-dbbe-4f74-a03c-72216a436e3e/7a80a5ff-f3bf-405b-902d-2c7a644821a3]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Creation complete after 20s [id=3e8dbe75-a73d-4282-b61b-1bdcc2b3c46d]",
                                "",
                                "Apply complete! Resources: 64 added, 0 changed, 0 destroyed.",
                                "",
                                "Outputs:",
                                "",
                                "manager_address = <sensitive>",
                                "private_key = <sensitive>"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000009-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:07.162672Z",
                            "start": "2026-03-13T00:02:26.739044Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000009",
                        "name": "Create infrastructure (latest)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_manager_version != \"latest\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:07.713013Z",
                            "start": "2026-03-13T00:04:07.202547Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000a",
                        "name": "Create infrastructure (stable)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTIuNDYK",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:08.401601Z",
                            "start": "2026-03-13T00:04:07.732239Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000b",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.192.46"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:08.549278Z",
                            "start": "2026-03-13T00:04:08.410819Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000c",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed03/src/github.com/osism/ansible-collection-commons"
                                    ],
                                    "delta": "0:00:00.699840",
                                    "end": "2026-03-13 00:04:09.561183",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed03/src/github.com/osism/ansible-collection-commons\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed03",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "commons",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-13 00:04:08.861343",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n[WARNING]: Collection osism.services does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.commons:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons'\nCreated collection for osism.commons:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons\nosism.commons:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.commons:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons'",
                                        "Created collection for osism.commons:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons",
                                        "osism.commons:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed03/src/github.com/osism/ansible-collection-services"
                                    ],
                                    "delta": "0:00:00.786279",
                                    "end": "2026-03-13 00:04:10.642825",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed03/src/github.com/osism/ansible-collection-services\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed03",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "services",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-03-13 00:04:09.856546",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.services does not support Ansible version 2.15.2\n[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.services:999.0.0' to '/home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services'",
                                        "Created collection for osism.services:999.0.0 at /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/services",
                                        "osism.services:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:10.705214Z",
                            "start": "2026-03-13T00:04:08.562722Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000d",
                        "name": "Update ansible collections"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 10,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.46",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:04:21.338100Z",
                            "start": "2026-03-13T00:04:10.717196Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000e",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-13 00:04:21.374976",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-13 00:05:21.378417",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:05:21.398760Z",
                            "start": "2026-03-13T00:04:21.349236Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000000f",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:05:22.986632Z",
                            "start": "2026-03-13T00:05:21.476182Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000010",
                        "name": "Fetch manager ssh hostkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "cp /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id* .",
                            "delta": "0:00:00.005431",
                            "end": "2026-03-13 00:05:23.482951",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "cp /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id* .",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000011-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:05:23.477520",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000011-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:05:23.547320Z",
                            "start": "2026-03-13T00:05:23.029807Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000011",
                        "name": "Get ssh keypair from terraform environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:05:23.607671Z",
                            "start": "2026-03-13T00:05:23.555484Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000012",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                            "delta": "0:10:09.286571",
                            "end": "2026-03-13 00:15:33.204007",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000013-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:05:23.917436",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Wait for cloud-init to finish] *******************************************\n\nTASK [Check /var/lib/cloud/instance/boot-finished] *****************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Update package cache] ****************************************************\nskipping: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nskipping: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Fail if Ubuntu version is lower than 24.04] ******************************\nskipping: [testbed-manager]\n\nTASK [Fail if Debian version is lower than 12] *********************************\nskipping: [testbed-manager]\n\nTASK [Set APT options on manager] **********************************************\nchanged: [testbed-manager]\n\nTASK [Update APT cache and run dist-upgrade] ***********************************\nchanged: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nchanged: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nchanged: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [Get current user] ********************************************************\nok: [testbed-manager]\n\nTASK [Create venv directory] ***************************************************\nchanged: [testbed-manager]\n\nTASK [Install netaddr in venv] *************************************************\nchanged: [testbed-manager]\n\nTASK [Install ansible-core in venv] ********************************************\nchanged: [testbed-manager]\n\nTASK [Install requests >= 2.32.2] **********************************************\nchanged: [testbed-manager]\n\nTASK [Install docker >= 7.1.0] *************************************************\nchanged: [testbed-manager]\n\nTASK [Create directories in /opt/src] ******************************************\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Sync sources in /opt/src] ************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Create /usr/share/ansible directory] *************************************\nchanged: [testbed-manager]\n\nTASK [Install collections from Ansible galaxy] *********************************\nchanged: [testbed-manager] => (item=ansible.netcommon)\nchanged: [testbed-manager] => (item=ansible.posix)\nchanged: [testbed-manager] => (item=community.docker>=3.10.2)\n\nTASK [Install local collections] ***********************************************\nchanged: [testbed-manager] => (item=ansible-collection-commons)\nchanged: [testbed-manager] => (item=ansible-collection-services)\n\nPLAY [Create operator user] ****************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Create operator group] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Create user] ************************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nchanged: [testbed-manager] => (item=adm)\nchanged: [testbed-manager] => (item=sudo)\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nchanged: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)\nchanged: [testbed-manager] => (item=export LANG=C.UTF-8)\nchanged: [testbed-manager] => (item=export LC_ALL=C.UTF-8)\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set password] ***********************************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Recursively change ownership of /opt/venv] *******************************\nchanged: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Wait for cloud-init to finish] *******************************************",
                                "",
                                "TASK [Check /var/lib/cloud/instance/boot-finished] *****************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Update package cache] ****************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Ubuntu version is lower than 24.04] ******************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Debian version is lower than 12] *********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set APT options on manager] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Update APT cache and run dist-upgrade] ***********************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get current user] ********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Create venv directory] ***************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install netaddr in venv] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install ansible-core in venv] ********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install requests >= 2.32.2] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install docker >= 7.1.0] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create directories in /opt/src] ******************************************",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Sync sources in /opt/src] ************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Create /usr/share/ansible directory] *************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install collections from Ansible galaxy] *********************************",
                                "changed: [testbed-manager] => (item=ansible.netcommon)",
                                "changed: [testbed-manager] => (item=ansible.posix)",
                                "changed: [testbed-manager] => (item=community.docker>=3.10.2)",
                                "",
                                "TASK [Install local collections] ***********************************************",
                                "changed: [testbed-manager] => (item=ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=ansible-collection-services)",
                                "",
                                "PLAY [Create operator user] ****************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "changed: [testbed-manager] => (item=adm)",
                                "changed: [testbed-manager] => (item=sudo)",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LANG=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8)",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Recursively change ownership of /opt/venv] *******************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000013-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:15:33.570462Z",
                            "start": "2026-03-13T00:05:23.645736Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000013",
                        "name": "Run manager part 0"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "It is now already possible to log in to the manager with 'make login'."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:15:33.622593Z",
                            "start": "2026-03-13T00:15:33.585893Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000014",
                        "name": "Point out that the log in on the manager is now possible"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:15:33.677561Z",
                            "start": "2026-03-13T00:15:33.630681Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000015",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                            "delta": "0:02:17.277077",
                            "end": "2026-03-13 00:17:51.301121",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed03/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000016-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:15:34.024044",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Run manager part 1] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nincluded: /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nskipping: [testbed-manager]\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Update package cache] *************************\nchanged: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Copy SSH public key] *****************************************************\nchanged: [testbed-manager]\n\nTASK [Copy SSH private key] ****************************************************\nchanged: [testbed-manager]\n\nTASK [Create configuration directory] ******************************************\nchanged: [testbed-manager]\n\nTASK [Copy testbed repo] *******************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager]\n\nTASK [Install python requirements in venv] *************************************\nok: [testbed-manager] => (item=Jinja2)\nok: [testbed-manager] => (item=PyYAML)\nok: [testbed-manager] => (item=packaging)\nchanged: [testbed-manager] => (item=python-gilt==1.2.3)\nok: [testbed-manager] => (item=requests>=2.32.2)\nok: [testbed-manager] => (item=docker>=7.1.0)\n\nTASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************\nchanged: [testbed-manager]\n\nTASK [Copy testbed custom CA certificate on CentOS] ****************************\nskipping: [testbed-manager]\n\nTASK [Run update-ca-certificates on Debian/Ubuntu] *****************************\nchanged: [testbed-manager]\n\nTASK [Run update-ca-trust on RedHat] *******************************************\nskipping: [testbed-manager]\n\nTASK [Run manager part 2] ******************************************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nok: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Run manager part 1] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "included: /home/zuul-testbed03/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Copy SSH public key] *****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy SSH private key] ****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create configuration directory] ******************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed repo] *******************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install python requirements in venv] *************************************",
                                "ok: [testbed-manager] => (item=Jinja2)",
                                "ok: [testbed-manager] => (item=PyYAML)",
                                "ok: [testbed-manager] => (item=packaging)",
                                "changed: [testbed-manager] => (item=python-gilt==1.2.3)",
                                "ok: [testbed-manager] => (item=requests>=2.32.2)",
                                "ok: [testbed-manager] => (item=docker>=7.1.0)",
                                "",
                                "TASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed custom CA certificate on CentOS] ****************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-certificates on Debian/Ubuntu] *****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-trust on RedHat] *******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run manager part 2] ******************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000016-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:17:51.360183Z",
                            "start": "2026-03-13T00:15:33.711968Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000016",
                        "name": "Run manager part 1 + 2"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "sudo",
                                "shutdown",
                                "-r",
                                "now"
                            ],
                            "delta": "0:00:00.878395",
                            "end": "2026-03-13 00:17:52.561936",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 sudo shutdown -r now",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000017-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:17:51.683541",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-000000000017-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:17:52.909042Z",
                            "start": "2026-03-13T00:17:51.395740Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000017",
                        "name": "Reboot manager"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 13,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.46",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:18:06.343966Z",
                            "start": "2026-03-13T00:17:52.924474Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000018",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-03-13 00:18:06.376971",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-03-13 00:19:06.380079",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:19:06.405396Z",
                            "start": "2026-03-13T00:18:06.353105Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000019",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "/opt/configuration/scripts/deploy-manager.sh"
                            ],
                            "delta": "0:23:24.002975",
                            "end": "2026-03-13 00:42:31.123151",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 /opt/configuration/scripts/deploy-manager.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001a-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:19:07.120176",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY MANAGER\n\n+ set -e\n+ echo\n+ echo '# DEPLOY MANAGER'\n+ echo\n+ cat /opt/manager-vars.sh\nexport NUMBER_OF_NODES=6\n\nexport CEPH_VERSION=reef\nexport CONFIGURATION_VERSION=main\nexport MANAGER_VERSION=latest\nexport OPENSTACK_VERSION=2025.1\n\nexport ARA=false\nexport DEPLOY_MODE=manager\nexport TEMPEST=true\nexport IS_ZUUL=true\n\nexport MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\nexport EXTERNAL_API=false\n\nexport IMAGE_USER=ubuntu\nexport IMAGE_NODE_USER=ubuntu\n\nexport CEPH_STACK=ceph-ansible\n\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver\n+ docker version\nClient: Docker Engine - Community\n Version:           27.5.1\n API version:       1.47\n Go version:        go1.22.11\n Git commit:        9f9e405\n Built:             Wed Jan 22 13:41:48 2025\n OS/Arch:           linux/amd64\n Context:           default\n\nServer: Docker Engine - Community\n Engine:\n  Version:          27.5.1\n  API version:      1.47 (minimum version 1.24)\n  Go version:       go1.22.11\n  Git commit:       4c9b3b0\n  Built:            Wed Jan 22 13:41:48 2025\n  OS/Arch:          linux/amd64\n  Experimental:     false\n containerd:\n  Version:          v2.2.2\n  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9\n runc:\n  Version:          1.3.4\n  GitCommit:        v1.3.4-0-gd6d73eb8\n docker-init:\n  Version:          0.19.0\n  GitCommit:        de40ad0\n+ sh -c /opt/configuration/scripts/deploy/000-manager.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-ceph-version.sh reef\n+ set -e\n+ VERSION=reef\n++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n ceph_version: reef ]]\n+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml\n+ /opt/configuration/scripts/set-openstack-version.sh 2025.1\n+ set -e\n+ VERSION=2025.1\n++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n openstack_version: 2024.2 ]]\n+ sed -i 's/openstack_version: .*/openstack_version: 2025.1/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ ceph-ansible == \\r\\o\\o\\k ]]\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ echo 'enable_osism_kubernetes: true'\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n++ semver 2025.1 2025.1\n+ [[ 0 -ge 0 ]]\n+ sed -i '/^om_enable_rabbitmq_high_availability:/d' /opt/configuration/environments/kolla/configuration.yml\n+ sed -i '/^om_enable_rabbitmq_quorum_queues:/d' /opt/configuration/environments/kolla/configuration.yml\n+ /opt/configuration/scripts/enable-resource-nodes.sh\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml\n\nPLAY [Copy custom facts] *******************************************************\n\nTASK [Create custom facts directory] *******************************************\nok: [testbed-manager]\n\nTASK [Copy fact files] *********************************************************\nchanged: [testbed-manager]\n\nPLAY [Before the deployment of the manager] ************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get /opt/manager-vars.sh] ************************************************\nok: [testbed-manager]\n\nTASK [Add ara_server_mariadb_volume_type parameter] ****************************\nchanged: [testbed-manager]\n\nTASK [Add netbox_enable parameter] *********************************************\nskipping: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Check if /etc/OTC_region exist] ******************************************\nok: [testbed-manager]\n\nTASK [Add nova_compute_virt_type parameter] ************************************\nskipping: [testbed-manager]\n\nPLAY [Apply role traefik] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply traefik role] ******************************************************\nincluded: osism.services.traefik for testbed-manager\n\nTASK [osism.services.traefik : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager\n\nTASK [osism.services.traefik : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/traefik)\nchanged: [testbed-manager] => (item=/opt/traefik/certificates)\nchanged: [testbed-manager] => (item=/opt/traefik/configuration)\n\nTASK [osism.services.traefik : Copy configuration files] ***********************\nchanged: [testbed-manager] => (item=traefik.yml)\nchanged: [testbed-manager] => (item=traefik.env)\nchanged: [testbed-manager] => (item=certificates.yml)\n\nTASK [osism.services.traefik : Copy certificate cert files] ********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy certificate key files] *********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy dynamic configuration] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.traefik : Remove dynamic configuration] *******************\nok: [testbed-manager]\n\nTASK [osism.services.traefik : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager\n\nTASK [osism.services.traefik : Create traefik external network] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Manage traefik service] *************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.traefik : Restart traefik service] *************\nskipping: [testbed-manager]\n\nPLAY [Deploy manager service] **************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply manager role] ******************************************************\nincluded: osism.services.manager for testbed-manager\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/ansible)\nchanged: [testbed-manager] => (item=/opt/archive)\nchanged: [testbed-manager] => (item=/opt/manager/configuration)\nchanged: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nchanged: [testbed-manager] => (item=/opt/manager/secrets)\nchanged: [testbed-manager] => (item=/opt/ansible/secrets)\nchanged: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nchanged: [testbed-manager] => (item=ara)\nchanged: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nchanged: [testbed-manager] => (item=conductor)\nchanged: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nchanged: [testbed-manager] => (item=osism)\nchanged: [testbed-manager] => (item=osism-update-docker)\nchanged: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nchanged: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nok: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nskipping: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:2025.1\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:2025.1\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   \n\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:2025.1             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   \nmanager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago        Up About a minute (healthy)   192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago        Up About a minute (healthy)   8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago        Up About a minute (healthy)   \nmanager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago        Up About a minute (healthy)   \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 57 seconds (healthy)       \nmanager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago        Up About a minute (healthy)   \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago        Up About a minute (healthy)   3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago        Up About a minute (healthy)   \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago        Up About a minute (healthy)   6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   \nosism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago        Up About a minute             192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   \nosismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago        Up About a minute (healthy)   \n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg\n+ osism apply resolvconf -l testbed-manager\n\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m\n\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87d64233-f8c7-4af4-b070-9f08313cd5a9 (resolvconf) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87d64233-f8c7-4af4-b070-9f08313cd5a9 (resolvconf) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role resolvconf] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nFriday 13 March 2026  00:23:17 +0000 (0:00:00.110)       0:00:00.110 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nFriday 13 March 2026  00:23:20 +0000 (0:00:03.270)       0:00:03.380 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nFriday 13 March 2026  00:23:20 +0000 (0:00:00.053)       0:00:03.434 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nFriday 13 March 2026  00:23:20 +0000 (0:00:00.077)       0:00:03.512 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nFriday 13 March 2026  00:23:21 +0000 (0:00:00.069)       0:00:03.581 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nFriday 13 March 2026  00:23:21 +0000 (0:00:00.853)       0:00:04.435 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nFriday 13 March 2026  00:23:21 +0000 (0:00:00.066)       0:00:04.501 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nFriday 13 March 2026  00:23:22 +0000 (0:00:00.471)       0:00:04.972 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nFriday 13 March 2026  00:23:22 +0000 (0:00:00.079)       0:00:05.052 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nFriday 13 March 2026  00:23:23 +0000 (0:00:00.542)       0:00:05.594 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nFriday 13 March 2026  00:23:24 +0000 (0:00:01.026)       0:00:06.621 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nFriday 13 March 2026  00:23:26 +0000 (0:00:01.962)       0:00:08.583 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nFriday 13 March 2026  00:23:26 +0000 (0:00:00.080)       0:00:08.664 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:23:27 +0000 (0:00:01.116)       0:00:09.780 ********** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.27s\nosism.commons.resolvconf : Start/enable systemd-resolved service -------- 1.96s\nosism.commons.resolvconf : Restart systemd-resolved service ------------- 1.12s\nosism.commons.resolvconf : Copy configuration files --------------------- 1.03s\nosism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 0.85s\nosism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.54s\nosism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.47s\nosism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s\nosism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.08s\nosism.commons.resolvconf : Include resolvconf tasks --------------------- 0.08s\nosism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s\nosism.commons.resolvconf : Install package systemd-resolved ------------- 0.07s\nosism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.05s\n+ osism apply sshconfig\n\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m\n\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3d8db8b-09c0-4107-8621-a25ba374ce63 (sshconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f3d8db8b-09c0-4107-8621-a25ba374ce63 (sshconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role sshconfig] ****************************************************\n\nTASK [osism.commons.sshconfig : Get home directory of operator user] ***********\nFriday 13 March 2026  00:23:43 +0000 (0:00:00.160)       0:00:00.160 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************\nFriday 13 March 2026  00:23:44 +0000 (0:00:00.551)       0:00:00.711 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure config for each host exist] *************\nFriday 13 March 2026  00:23:45 +0000 (0:00:00.498)       0:00:01.210 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\nTASK [osism.commons.sshconfig : Add extra config] ******************************\nFriday 13 March 2026  00:23:50 +0000 (0:00:05.494)       0:00:06.705 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Assemble ssh config] ***************************\nFriday 13 March 2026  00:23:50 +0000 (0:00:00.057)       0:00:06.762 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:23:51 +0000 (0:00:00.525)       0:00:07.288 ********** \n=============================================================================== \nosism.commons.sshconfig : Ensure config for each host exist ------------- 5.49s\nosism.commons.sshconfig : Get home directory of operator user ----------- 0.55s\nosism.commons.sshconfig : Assemble ssh config --------------------------- 0.53s\nosism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.50s\nosism.commons.sshconfig : Add extra config ------------------------------ 0.06s\n+ osism apply known-hosts\n\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m\n\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9eb0a440-9b1a-4fed-95c7-27da1b6b3a51 (known-hosts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9eb0a440-9b1a-4fed-95c7-27da1b6b3a51 (known-hosts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role known_hosts] **************************************************\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***\nFriday 13 March 2026  00:24:07 +0000 (0:00:00.161)       0:00:00.161 ********** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***\nFriday 13 March 2026  00:24:13 +0000 (0:00:05.868)       0:00:06.029 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:13 +0000 (0:00:00.164)       0:00:06.194 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJuy8jCWnaLTBq9iEMruLjIH5rb6Sdsk/mdqEoBrT+Yh)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDUjn3pbkhTVDSfIvFWyrTPSAlrQ2qEHx+cKiucmRnylN8/dOPmVnzIWXKglSPdWPk0RXKjrAqkgk3asrjIyONAMu+Yodj9dMm7cQyJugykZzlsxrrqf29B1jllI0Vv0CmG+0aNOeySPU9lAjyCBT5bsvRHvFy6snz+unXWZ2P7ptsbZe5u7K5Gpl+5zsQpESQ/mvU2X1Rtvgn+7kxrwa0meG+JS/WYnluFbB1UBpYaaynx+6afbQBzJmSHJPU1PbbcVumyIvkBj2XL4ivi7Hi4HDtjSyICV22yl2/gSe6TrqPpRU/pAIy+w2eBb02FPcrt0fPb4UXUN1wqFPBSXYpA2UA89dw/lHJSTZNfCggEAEWXL3QGap2HTSiDliiowmuogw3hXL3l/M4RmA+CWB9V/sLY9ziYWc5p+3PBYs7SZCCQUy0nRFs82CqxQdBUnYu5ZJKZ8Jvo2ciXijV47faOxkEKDV5ur6sIYUuHJ3Cs5tPhSLFNL4l5Go2owYA8Sok=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCa5ZI05+fptRQNYqGW5ymCBzvxotWxdg4u9MPllpnlgiAaEzku1jQXOnCFnRCqR4no9Bp6IP/XbXsA6cIlx+cg=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:14 +0000 (0:00:01.164)       0:00:07.358 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBlWLCAhw75OFUH+WfJRsWoE7r4lRId0a5SuaQcF3RCn77R6tGXubcwvekH3i36AYoPVZ1f28/in+9qmOfBZEdg=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCx2OXD5KvuxQQvuuk/ZR3TYhjd2oCUgvGGlLl59UdbEWb6avU9P2aOZn394QvLxl/W0O5CVw+zcEx4rci/P5nYCHHzHb2uro0bRaBfdqQ0B4uqjFCIO8/aIULyDiBg0pDyxW6GJI0qERaXhOzc8z9bbyh9ve7zxqRm4M8cexuc+/79N/TqmUTbg+vQTTGJrHcKQrHFxIuBoH6NLeQaLRXfl6mb+2NmKp1bHHnrB0/QCBs6ry0PyJAgH6cLlD4sAiU8vC4IAkOVJu8tH6xA5ADISAS/SYx7M0F2+Ai9zUg3m+z3Kx0Ye9ZX+fyIeBm+dF9BV4/w6sSGqA/7L92pthy1dvpMADYsbNGL2graufYaxXiDCy4sGIzzPnzQCimpmIPkB7btUi/Fg6lzIG+RgHEyabKa++kbUPALty/qKm/ZLWsEwzoIvpMIZ8thOCHmYjHAYixR0kJO50v+A4udUDjmolwsMz8F9RCtxvxdgbHCN0MPKHNTNFENF/MPjLtOgXE=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINFXmFixrYsGuD6iK/5RZVMg1Bk/21IF1SHB4JPBI/fc)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:15 +0000 (0:00:00.975)       0:00:08.333 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ893j22t3cuMmiIB4zlWJMjyKFbTwJcc0BPH2rFyRTcNsdUzljqJpt4If0YzK+kUx9KSF9+7IQYQ5r+GlO4w+HloWfyBBsrzV/lf1w3dB5vNG9fhbRwOfegY+OI51PhSsJz2aZcmgI9X7Dy7M/9OQs9qQFxwjdx4OGwhDu0j+MwYiAUMLmju52inCVoEte9gr6XrFDex3kQA9oZVsW886Milfpj7a/kI31O85i3+RkrS1zWa064cZmnN/vIiJnhjezoSIGD/RHqIJS/k3GCnJTNMX+4WNRbjWPoALbiSXZl/WnofheFKuV9MSa+ylX1xIk1ua1E4cgCzI1m68m3909ufoY2T0YIQahM9iifMT664yLltMZqtujwRdyqbKL1+Xh6Fu+0REtnZgi50dHRR0xDzYfGTkjGab7V61ma2lhvMtpDg389lQwXFnUeKgpYZ2TfLIkdmu6c0q5rwPkrPk2jl3hJ6yjXLlKLtZ/xVL6FLrgd+n3xXkWpEUNm66WWk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIO25puKqGyzZ0nxVExnHcaYjmw7grNnYJGgIjtxandl1sKw4yHddITZ96S2Xx34RmTga6QLQnkmN6cp3DjcIYw=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEx3YF+TLygLxtcQFHnbHVeSvusJcGQANr1LDnZFBER1)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:16 +0000 (0:00:01.014)       0:00:09.347 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJyptwywAKmXERvxIbipTGF0jGA3OcJIXuRaNN4oaasRFcHs6TIsI+eCRt0U0JeTo2gQls1444wdlWE0zcBq5hw=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDj8+jynPJOfpKxwEdGVPOTygQAdAZ45l3zF2eb0Z7tXue7wdHzMCuQbVDXN1e6hHJYWw3WLeOiNkdZ1jQRvGehGRX9mI0GV+g4Pxx3t/fxT7IOxiw5Vmc1w6uB/vWWEum2fTT//tmBToA+ZQK8BzbjY9xum4PsEz+frApichK3SkheQJYJFKSinttaw39//rEMu9f6ZZaKBKmQFRnWANQnEaBUsAGP8BrOHJnc9JorlecS0qnVKDXLjX36qOTUj7CG89fnZGx4hmra6BWUtqGMLbiZmmKADImxzyC1NMvlYoECnt2vepEC+vK8dV0ITuCTaGGvlYrDRFOFaBn+mWGcwqLsMy1uLzKk3eKB8hZJ/6lh8AqO6oygvJU/tBDpbadFVBxHRq8qXyoe0Lrtnncd4jUEcx4SrBzHO+HjKnStSHxY93VtYRDLsP6rc4utNl3wKk1VbXHzZ7NdssFE/+uznzPcI18xp7knIn8BCIGkIUd8gNJtw+XUiEdgdd9OARM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPXUTKHySIJuePlD23IvyWC7WWIcXkfHppc5thZrV4EL)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:17 +0000 (0:00:01.014)       0:00:10.362 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVi8GbWLEVnN2cq1UtQFymBGqLAadzfhhMKJN4hSR6qu1xpVVoC8m4wmpU/M0ryTlbp8HiNhe/5fHTQiRkite0mThpP9gcYDkXsJsJyIWhSZ+w9d1KdZy33iHVLFmgHWhTmCa3Zz9ptJDVZWyTSqIf9FV5RQJDVO4F43bf4Oz/ZKIKZGV+pb81Nl21s+OCcr3PnCchaKMgRHuUiz1wlaaE5gJeyS3muh1jpc9jxi47IAqlvbdWsP0n2iFFy2Vz+FSd+VcRoHL/4EnbAUMs8278wJlL8bcpGWH3qls69gc8vQRsUmH7HgZn8yC9CHTazbShsy8y3eXWOXyiDPQnmDlJVDr1fAWnBPVCqfo3gWvURBXbG8/MdLKfZmCDpYouL9T8UR6Y2SdQXtGFpx0L4Hb7Oloi8EiUt1d8k86Tqol7MxKL1v3VBct+8WVQUrwhtApgYOLT2XXSwuMb6wiKzuLswvyUOZzMoXipWboEoUl74mwoe17QSROul5/otMlzB08=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJcatKxhtat8HayXurSHkcdRyxEd3QAmE1LHT07qM0cl)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAW1cyPz5B71lNK9YON6FQwdjfZdu5EuN9ADHhJNE8uj72sCkjktOPZrPm9burSJUqBE9/+Ran1oU0b/+jtRqpA=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:18 +0000 (0:00:01.007)       0:00:11.370 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG+ZNRyzVJ0g92UyubSMydcqVN3BYyGRL3kj7QhcmgiEjpPyLxZ9nUsQ5jNAYVIdFk2aW7qkU7iitX5aRU9djz4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCrdcb4tS416Fq7fSiHkjU70NV58/hXHf3StBiWpNG/pF4gjnQjbaj31jEUY9Os6rniNll/iQ7pI4HG9pnZ3UOWjg7hk9VBz+lod9Fb8RrA4C4/WdJo20YR7u5RqlaQBSMnZksbvzaZCOdxqDF6BDNg3Eq1x149dqcyfCaoQinLS2zAHPZofL6VuvRib33wscojz7mZZQkgMfpty2ensqfW6jypAtaD1fz87wUb2Eu7c1oedv8DQlaYPCoAPWmdIuLpmcxAyK0jFYVf51mrWFeE3KV4wKDLwRU8fm8B4ciUFP8QkOXH70vx3NT2nc2zORhJAI9gqYWBg3LS+LGZL0Zm3cCy0JmyKeP6iN0X9Vz11T0EL1IkQj62xN6y6yG1ArZ46gNDfY5997YL7PbxGtD5840z0Ap2A8JPtu7eAOUuRfvAoDrOLdDqb6glRYeCMMKWEgpSFRdSIfoe/S4dx/3eDbZmm2ajcm/B8cO0fKiu9n96jP0hSQNcwtQUcl0drG8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN5nXX6lrkdkqdrcnFbDBZcznK46Ot6IznDrG1/JbJtR)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:19 +0000 (0:00:00.994)       0:00:12.365 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEq41I4ETswmUujQlL1aob78tYUYDVmH1bCbKnsi1rza)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBVzQyOT4WunW9bbwQfaHK8PAR6yp3J9G8oqLno+QCnGild5O03duxDMS2WWW+12cOjjlipXQEOWeckreVtIjBOlvfLkRpAbalcOHRAYO2ZD6+IhLX4zsCc+b5N+j6uGSjOlG2P/i9LIRJgyKeZoPdQnxRhQWmY9CP+duDs5jutmOZuzSDimO/S0gHdCqPJ+c9FsxX7AA3gZFBTjEqvHB8SA+hMKu3GCnDbhmx1FourmBF0E+2EQONrNazFmHZHEP8IxrLk4jRIwQ37N2Opoay913z0w4YvwCkcU1oh+fj0j8ggSwfKB0s62HA/qN9/p6z315oHJQYoJyr8+5+oyIS0piVUYORJrOIcpMdnIWepAJILND6E8OhX5YqFf06+nmf8y33tBFFJMziN52jm3fJZAeFR5vtVw4Uy24tixMHYY8IETvBrWxBwO13wDmeAAORfV7KHYQgdcwWbZlHDe1xnFNJ9v95FVaTuH2q7ceq4oAMFfnXRWRJgVxQH37WOXc=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMClF11FAZYr5fRF50WO1AWC8MiLmVAfUBX2NUtUP9tf987+Msed8myo78FC7NXXIO10EyVDIODNmrOuQXiJJ3A=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***\nFriday 13 March 2026  00:24:20 +0000 (0:00:00.978)       0:00:13.343 ********** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***\nFriday 13 March 2026  00:24:25 +0000 (0:00:05.230)       0:00:18.574 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:26 +0000 (0:00:00.171)       0:00:18.745 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJuy8jCWnaLTBq9iEMruLjIH5rb6Sdsk/mdqEoBrT+Yh)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDUjn3pbkhTVDSfIvFWyrTPSAlrQ2qEHx+cKiucmRnylN8/dOPmVnzIWXKglSPdWPk0RXKjrAqkgk3asrjIyONAMu+Yodj9dMm7cQyJugykZzlsxrrqf29B1jllI0Vv0CmG+0aNOeySPU9lAjyCBT5bsvRHvFy6snz+unXWZ2P7ptsbZe5u7K5Gpl+5zsQpESQ/mvU2X1Rtvgn+7kxrwa0meG+JS/WYnluFbB1UBpYaaynx+6afbQBzJmSHJPU1PbbcVumyIvkBj2XL4ivi7Hi4HDtjSyICV22yl2/gSe6TrqPpRU/pAIy+w2eBb02FPcrt0fPb4UXUN1wqFPBSXYpA2UA89dw/lHJSTZNfCggEAEWXL3QGap2HTSiDliiowmuogw3hXL3l/M4RmA+CWB9V/sLY9ziYWc5p+3PBYs7SZCCQUy0nRFs82CqxQdBUnYu5ZJKZ8Jvo2ciXijV47faOxkEKDV5ur6sIYUuHJ3Cs5tPhSLFNL4l5Go2owYA8Sok=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCa5ZI05+fptRQNYqGW5ymCBzvxotWxdg4u9MPllpnlgiAaEzku1jQXOnCFnRCqR4no9Bp6IP/XbXsA6cIlx+cg=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:27 +0000 (0:00:01.024)       0:00:19.769 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINFXmFixrYsGuD6iK/5RZVMg1Bk/21IF1SHB4JPBI/fc)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCx2OXD5KvuxQQvuuk/ZR3TYhjd2oCUgvGGlLl59UdbEWb6avU9P2aOZn394QvLxl/W0O5CVw+zcEx4rci/P5nYCHHzHb2uro0bRaBfdqQ0B4uqjFCIO8/aIULyDiBg0pDyxW6GJI0qERaXhOzc8z9bbyh9ve7zxqRm4M8cexuc+/79N/TqmUTbg+vQTTGJrHcKQrHFxIuBoH6NLeQaLRXfl6mb+2NmKp1bHHnrB0/QCBs6ry0PyJAgH6cLlD4sAiU8vC4IAkOVJu8tH6xA5ADISAS/SYx7M0F2+Ai9zUg3m+z3Kx0Ye9ZX+fyIeBm+dF9BV4/w6sSGqA/7L92pthy1dvpMADYsbNGL2graufYaxXiDCy4sGIzzPnzQCimpmIPkB7btUi/Fg6lzIG+RgHEyabKa++kbUPALty/qKm/ZLWsEwzoIvpMIZ8thOCHmYjHAYixR0kJO50v+A4udUDjmolwsMz8F9RCtxvxdgbHCN0MPKHNTNFENF/MPjLtOgXE=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBlWLCAhw75OFUH+WfJRsWoE7r4lRId0a5SuaQcF3RCn77R6tGXubcwvekH3i36AYoPVZ1f28/in+9qmOfBZEdg=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:28 +0000 (0:00:01.030)       0:00:20.800 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEx3YF+TLygLxtcQFHnbHVeSvusJcGQANr1LDnZFBER1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ893j22t3cuMmiIB4zlWJMjyKFbTwJcc0BPH2rFyRTcNsdUzljqJpt4If0YzK+kUx9KSF9+7IQYQ5r+GlO4w+HloWfyBBsrzV/lf1w3dB5vNG9fhbRwOfegY+OI51PhSsJz2aZcmgI9X7Dy7M/9OQs9qQFxwjdx4OGwhDu0j+MwYiAUMLmju52inCVoEte9gr6XrFDex3kQA9oZVsW886Milfpj7a/kI31O85i3+RkrS1zWa064cZmnN/vIiJnhjezoSIGD/RHqIJS/k3GCnJTNMX+4WNRbjWPoALbiSXZl/WnofheFKuV9MSa+ylX1xIk1ua1E4cgCzI1m68m3909ufoY2T0YIQahM9iifMT664yLltMZqtujwRdyqbKL1+Xh6Fu+0REtnZgi50dHRR0xDzYfGTkjGab7V61ma2lhvMtpDg389lQwXFnUeKgpYZ2TfLIkdmu6c0q5rwPkrPk2jl3hJ6yjXLlKLtZ/xVL6FLrgd+n3xXkWpEUNm66WWk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIO25puKqGyzZ0nxVExnHcaYjmw7grNnYJGgIjtxandl1sKw4yHddITZ96S2Xx34RmTga6QLQnkmN6cp3DjcIYw=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:29 +0000 (0:00:00.983)       0:00:21.784 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDj8+jynPJOfpKxwEdGVPOTygQAdAZ45l3zF2eb0Z7tXue7wdHzMCuQbVDXN1e6hHJYWw3WLeOiNkdZ1jQRvGehGRX9mI0GV+g4Pxx3t/fxT7IOxiw5Vmc1w6uB/vWWEum2fTT//tmBToA+ZQK8BzbjY9xum4PsEz+frApichK3SkheQJYJFKSinttaw39//rEMu9f6ZZaKBKmQFRnWANQnEaBUsAGP8BrOHJnc9JorlecS0qnVKDXLjX36qOTUj7CG89fnZGx4hmra6BWUtqGMLbiZmmKADImxzyC1NMvlYoECnt2vepEC+vK8dV0ITuCTaGGvlYrDRFOFaBn+mWGcwqLsMy1uLzKk3eKB8hZJ/6lh8AqO6oygvJU/tBDpbadFVBxHRq8qXyoe0Lrtnncd4jUEcx4SrBzHO+HjKnStSHxY93VtYRDLsP6rc4utNl3wKk1VbXHzZ7NdssFE/+uznzPcI18xp7knIn8BCIGkIUd8gNJtw+XUiEdgdd9OARM=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJyptwywAKmXERvxIbipTGF0jGA3OcJIXuRaNN4oaasRFcHs6TIsI+eCRt0U0JeTo2gQls1444wdlWE0zcBq5hw=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPXUTKHySIJuePlD23IvyWC7WWIcXkfHppc5thZrV4EL)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:30 +0000 (0:00:01.006)       0:00:22.790 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAW1cyPz5B71lNK9YON6FQwdjfZdu5EuN9ADHhJNE8uj72sCkjktOPZrPm9burSJUqBE9/+Ran1oU0b/+jtRqpA=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVi8GbWLEVnN2cq1UtQFymBGqLAadzfhhMKJN4hSR6qu1xpVVoC8m4wmpU/M0ryTlbp8HiNhe/5fHTQiRkite0mThpP9gcYDkXsJsJyIWhSZ+w9d1KdZy33iHVLFmgHWhTmCa3Zz9ptJDVZWyTSqIf9FV5RQJDVO4F43bf4Oz/ZKIKZGV+pb81Nl21s+OCcr3PnCchaKMgRHuUiz1wlaaE5gJeyS3muh1jpc9jxi47IAqlvbdWsP0n2iFFy2Vz+FSd+VcRoHL/4EnbAUMs8278wJlL8bcpGWH3qls69gc8vQRsUmH7HgZn8yC9CHTazbShsy8y3eXWOXyiDPQnmDlJVDr1fAWnBPVCqfo3gWvURBXbG8/MdLKfZmCDpYouL9T8UR6Y2SdQXtGFpx0L4Hb7Oloi8EiUt1d8k86Tqol7MxKL1v3VBct+8WVQUrwhtApgYOLT2XXSwuMb6wiKzuLswvyUOZzMoXipWboEoUl74mwoe17QSROul5/otMlzB08=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJcatKxhtat8HayXurSHkcdRyxEd3QAmE1LHT07qM0cl)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:31 +0000 (0:00:00.974)       0:00:23.765 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG+ZNRyzVJ0g92UyubSMydcqVN3BYyGRL3kj7QhcmgiEjpPyLxZ9nUsQ5jNAYVIdFk2aW7qkU7iitX5aRU9djz4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCrdcb4tS416Fq7fSiHkjU70NV58/hXHf3StBiWpNG/pF4gjnQjbaj31jEUY9Os6rniNll/iQ7pI4HG9pnZ3UOWjg7hk9VBz+lod9Fb8RrA4C4/WdJo20YR7u5RqlaQBSMnZksbvzaZCOdxqDF6BDNg3Eq1x149dqcyfCaoQinLS2zAHPZofL6VuvRib33wscojz7mZZQkgMfpty2ensqfW6jypAtaD1fz87wUb2Eu7c1oedv8DQlaYPCoAPWmdIuLpmcxAyK0jFYVf51mrWFeE3KV4wKDLwRU8fm8B4ciUFP8QkOXH70vx3NT2nc2zORhJAI9gqYWBg3LS+LGZL0Zm3cCy0JmyKeP6iN0X9Vz11T0EL1IkQj62xN6y6yG1ArZ46gNDfY5997YL7PbxGtD5840z0Ap2A8JPtu7eAOUuRfvAoDrOLdDqb6glRYeCMMKWEgpSFRdSIfoe/S4dx/3eDbZmm2ajcm/B8cO0fKiu9n96jP0hSQNcwtQUcl0drG8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN5nXX6lrkdkqdrcnFbDBZcznK46Ot6IznDrG1/JbJtR)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nFriday 13 March 2026  00:24:32 +0000 (0:00:01.008)       0:00:24.773 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEq41I4ETswmUujQlL1aob78tYUYDVmH1bCbKnsi1rza)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBVzQyOT4WunW9bbwQfaHK8PAR6yp3J9G8oqLno+QCnGild5O03duxDMS2WWW+12cOjjlipXQEOWeckreVtIjBOlvfLkRpAbalcOHRAYO2ZD6+IhLX4zsCc+b5N+j6uGSjOlG2P/i9LIRJgyKeZoPdQnxRhQWmY9CP+duDs5jutmOZuzSDimO/S0gHdCqPJ+c9FsxX7AA3gZFBTjEqvHB8SA+hMKu3GCnDbhmx1FourmBF0E+2EQONrNazFmHZHEP8IxrLk4jRIwQ37N2Opoay913z0w4YvwCkcU1oh+fj0j8ggSwfKB0s62HA/qN9/p6z315oHJQYoJyr8+5+oyIS0piVUYORJrOIcpMdnIWepAJILND6E8OhX5YqFf06+nmf8y33tBFFJMziN52jm3fJZAeFR5vtVw4Uy24tixMHYY8IETvBrWxBwO13wDmeAAORfV7KHYQgdcwWbZlHDe1xnFNJ9v95FVaTuH2q7ceq4oAMFfnXRWRJgVxQH37WOXc=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMClF11FAZYr5fRF50WO1AWC8MiLmVAfUBX2NUtUP9tf987+Msed8myo78FC7NXXIO10EyVDIODNmrOuQXiJJ3A=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write static known_hosts entries] ************\nFriday 13 March 2026  00:24:33 +0000 (0:00:01.050)       0:00:25.824 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Write extra known_hosts entries] *************\nFriday 13 March 2026  00:24:33 +0000 (0:00:00.155)       0:00:25.979 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Delete known_hosts entries] ******************\nFriday 13 March 2026  00:24:33 +0000 (0:00:00.046)       0:00:26.026 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Set file permissions] ************************\nFriday 13 March 2026  00:24:33 +0000 (0:00:00.041)       0:00:26.067 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:24:34 +0000 (0:00:00.701)       0:00:26.768 ********** \n=============================================================================== \nosism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.87s\nosism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.23s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.16s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.99s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s\nosism.commons.known_hosts : Set file permissions ------------------------ 0.70s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s\nosism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s\n+ osism apply squid\n\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m\n\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 076722ab-df37-4595-91ef-2a9c0563a707 (squid) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 076722ab-df37-4595-91ef-2a9c0563a707 (squid) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role squid] ********************************************************\n\nTASK [osism.services.squid : Include install tasks] ****************************\nFriday 13 March 2026  00:24:50 +0000 (0:00:00.119)       0:00:00.119 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.squid : Install required packages] ************************\nFriday 13 March 2026  00:24:50 +0000 (0:00:00.063)       0:00:00.183 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Create required directories] **********************\nFriday 13 March 2026  00:24:51 +0000 (0:00:01.011)       0:00:01.194 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m\n\nTASK [osism.services.squid : Copy squid configuration files] *******************\nFriday 13 March 2026  00:24:52 +0000 (0:00:01.004)       0:00:02.199 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m\n\nTASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***\nFriday 13 March 2026  00:24:53 +0000 (0:00:00.953)       0:00:03.153 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Copy docker-compose.yml file] *********************\nFriday 13 March 2026  00:24:53 +0000 (0:00:00.318)       0:00:03.471 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Manage squid service] *****************************\nFriday 13 March 2026  00:24:54 +0000 (0:00:00.789)       0:00:04.261 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Restart squid service] *****************\nFriday 13 March 2026  00:25:24 +0000 (0:00:29.893)       0:00:34.155 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******\nFriday 13 March 2026  00:25:36 +0000 (0:00:11.926)       0:00:46.081 ********** \nPausing for 60 seconds\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***\nFriday 13 March 2026  00:26:36 +0000 (0:01:00.084)       0:01:46.165 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****\nFriday 13 March 2026  00:26:36 +0000 (0:00:00.059)       0:01:46.224 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:26:36 +0000 (0:00:00.597)       0:01:46.821 ********** \n=============================================================================== \nosism.services.squid : Wait for squid service to start ----------------- 60.08s\nosism.services.squid : Manage squid service ---------------------------- 29.89s\nosism.services.squid : Restart squid service --------------------------- 11.93s\nosism.services.squid : Install required packages ------------------------ 1.01s\nosism.services.squid : Create required directories ---------------------- 1.00s\nosism.services.squid : Copy squid configuration files ------------------- 0.95s\nosism.services.squid : Copy docker-compose.yml file --------------------- 0.79s\nosism.services.squid : Wait for an healthy squid service ---------------- 0.60s\nosism.services.squid : Remove osism_allow_list.conf configuration file --- 0.32s\nosism.services.squid : Include install tasks ---------------------------- 0.06s\nosism.services.squid : Register that squid service was restarted -------- 0.06s\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla\n+ set -e\n+ NAMESPACE=kolla\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n++ semver latest 9.0.0\n+ [[ -1 -lt 0 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ osism apply operator -u ubuntu -l testbed-nodes\n\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m\n\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a138e641-c2e4-42c8-867d-347889857868 (operator) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a138e641-c2e4-42c8-867d-347889857868 (operator) has been started and output is visible here.\u001b[0m\n\nPLAY [Make ssh pipelining working] *********************************************\n\nTASK [Gathering Facts] *********************************************************\nFriday 13 March 2026  00:26:53 +0000 (0:00:00.101)       0:00:00.101 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Do not require tty for all users] ****************************************\nFriday 13 March 2026  00:26:57 +0000 (0:00:04.234)       0:00:04.336 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role operator] *****************************************************\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nFriday 13 March 2026  00:26:58 +0000 (0:00:00.948)       0:00:05.285 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nFriday 13 March 2026  00:26:58 +0000 (0:00:00.145)       0:00:05.431 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create operator group] **************************\nFriday 13 March 2026  00:26:58 +0000 (0:00:00.165)       0:00:05.596 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.operator : Create user] ************************************\nFriday 13 March 2026  00:26:59 +0000 (0:00:00.645)       0:00:06.242 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nFriday 13 March 2026  00:27:00 +0000 (0:00:00.953)       0:00:07.195 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nFriday 13 March 2026  00:27:01 +0000 (0:00:01.242)       0:00:08.437 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nFriday 13 March 2026  00:27:02 +0000 (0:00:01.323)       0:00:09.760 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m\n\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m\n\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nFriday 13 March 2026  00:27:04 +0000 (0:00:01.280)       0:00:11.042 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nFriday 13 March 2026  00:27:04 +0000 (0:00:00.134)       0:00:11.176 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nFriday 13 March 2026  00:27:04 +0000 (0:00:00.165)       0:00:11.342 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nFriday 13 March 2026  00:27:05 +0000 (0:00:00.656)       0:00:11.998 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nFriday 13 March 2026  00:27:05 +0000 (0:00:00.164)       0:00:12.163 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nFriday 13 March 2026  00:27:05 +0000 (0:00:00.731)       0:00:12.894 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nFriday 13 March 2026  00:27:06 +0000 (0:00:00.129)       0:00:13.024 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nFriday 13 March 2026  00:27:06 +0000 (0:00:00.145)       0:00:13.169 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set password] ***********************************\nFriday 13 March 2026  00:27:06 +0000 (0:00:00.141)       0:00:13.311 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nFriday 13 March 2026  00:27:07 +0000 (0:00:00.701)       0:00:14.012 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:27:07 +0000 (0:00:00.214)       0:00:14.226 ********** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 4.23s\nosism.commons.operator : Copy user sudoers file ------------------------- 1.32s\nosism.commons.operator : Set language variables in .bashrc configuration file --- 1.28s\nosism.commons.operator : Add user to additional groups ------------------ 1.24s\nosism.commons.operator : Create user ------------------------------------ 0.95s\nDo not require tty for all users ---------------------------------------- 0.95s\nosism.commons.operator : Set ssh authorized keys ------------------------ 0.73s\nosism.commons.operator : Set password ----------------------------------- 0.70s\nosism.commons.operator : Create .ssh directory -------------------------- 0.66s\nosism.commons.operator : Create operator group -------------------------- 0.65s\nosism.commons.operator : Unset & lock password -------------------------- 0.21s\nosism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.17s\nosism.commons.operator : Set operator_groups variable to default value --- 0.17s\nosism.commons.operator : Check number of SSH authorized keys ------------ 0.16s\nosism.commons.operator : Gather variables for each operating system ----- 0.15s\nosism.commons.operator : Set authorized GitHub accounts ----------------- 0.15s\nosism.commons.operator : Delete authorized GitHub accounts -------------- 0.14s\nosism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.13s\nosism.commons.operator : Delete ssh authorized keys --------------------- 0.13s\n+ osism apply --environment custom facts\n\u001b[32m2026-03-13 00:27:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m\n\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7bcd12de-b383-4c91-a9da-2851d605c017 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7bcd12de-b383-4c91-a9da-2851d605c017 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy custom network devices fact] ****************************************\n\nTASK [Create custom facts directory] *******************************************\nFriday 13 March 2026  00:27:23 +0000 (0:00:00.052)       0:00:00.052 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Copy fact file] **********************************************************\nFriday 13 March 2026  00:27:24 +0000 (0:00:01.255)       0:00:01.307 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Copy custom ceph devices facts] ******************************************\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nFriday 13 March 2026  00:27:25 +0000 (0:00:01.018)       0:00:02.326 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nFriday 13 March 2026  00:27:25 +0000 (0:00:00.087)       0:00:02.414 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nFriday 13 March 2026  00:27:25 +0000 (0:00:00.164)       0:00:02.578 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nFriday 13 March 2026  00:27:25 +0000 (0:00:00.177)       0:00:02.756 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nFriday 13 March 2026  00:27:26 +0000 (0:00:00.097)       0:00:02.853 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nFriday 13 March 2026  00:27:26 +0000 (0:00:00.439)       0:00:03.293 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nFriday 13 March 2026  00:27:26 +0000 (0:00:00.086)       0:00:03.379 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nFriday 13 March 2026  00:27:27 +0000 (0:00:01.005)       0:00:04.384 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nFriday 13 March 2026  00:27:28 +0000 (0:00:00.438)       0:00:04.823 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nFriday 13 March 2026  00:27:29 +0000 (0:00:01.086)       0:00:05.909 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Install required packages (RedHat)] **************************************\nFriday 13 March 2026  00:27:46 +0000 (0:00:17.755)       0:00:23.664 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (Debian)] **************************************\nFriday 13 March 2026  00:27:46 +0000 (0:00:00.103)       0:00:23.768 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Create custom facts directory] *******************************************\nFriday 13 March 2026  00:27:56 +0000 (0:00:09.147)       0:00:32.915 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Copy fact files] *********************************************************\nFriday 13 March 2026  00:27:56 +0000 (0:00:00.442)       0:00:33.358 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nFriday 13 March 2026  00:28:00 +0000 (0:00:03.739)       0:00:37.097 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:28:01 +0000 (0:00:01.587)       0:00:38.685 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:28:06 +0000 (0:00:04.840)       0:00:43.525 ********** \n=============================================================================== \nosism.commons.repository : Update package cache ------------------------ 17.76s\nInstall required packages (Debian) -------------------------------------- 9.15s\nGathers facts about hosts ----------------------------------------------- 4.84s\nCopy fact files --------------------------------------------------------- 3.74s\nosism.commons.repository : Force update of package cache ---------------- 1.59s\nCreate custom facts directory ------------------------------------------- 1.26s\nosism.commons.repository : Copy ubuntu.sources file --------------------- 1.09s\nCopy fact file ---------------------------------------------------------- 1.02s\nosism.commons.repository : Copy 99osism apt configuration --------------- 1.01s\nCreate custom facts directory ------------------------------------------- 0.44s\nosism.commons.repository : Remove sources.list file --------------------- 0.44s\nosism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.44s\nosism.commons.repository : Set repositories to default ------------------ 0.18s\nosism.commons.repository : Set repository_default fact to default value --- 0.16s\nInstall required packages (RedHat) -------------------------------------- 0.10s\nosism.commons.repository : Include distribution specific repository tasks --- 0.10s\nosism.commons.repository : Gather variables for each operating system --- 0.09s\nosism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.09s\n+ osism apply bootstrap\n\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m\n\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62f251aa-bb4c-40fe-8fe8-9c5aca7a7d34 (bootstrap) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62f251aa-bb4c-40fe-8fe8-9c5aca7a7d34 (bootstrap) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on state bootstrap] ************************************\n\nTASK [Group hosts based on state bootstrap] ************************************\nFriday 13 March 2026  00:28:23 +0000 (0:00:00.137)       0:00:00.137 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:28:23 +0000 (0:00:00.229)       0:00:00.367 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY [Gather facts for all hosts (if using --limit)] ***************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:28:27 +0000 (0:00:03.546)       0:00:03.914 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Apply bootstrap roles part 1] ********************************************\n\nTASK [osism.commons.hostname : Set hostname] ***********************************\nFriday 13 March 2026  00:28:27 +0000 (0:00:00.491)       0:00:04.405 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.hostname : Copy /etc/hostname] *****************************\nFriday 13 March 2026  00:28:29 +0000 (0:00:01.257)       0:00:05.663 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.hosts : Include type specific tasks] ***********************\nFriday 13 March 2026  00:28:30 +0000 (0:00:01.401)       0:00:07.065 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.hosts : Copy /etc/hosts file] ******************************\nFriday 13 March 2026  00:28:30 +0000 (0:00:00.285)       0:00:07.351 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.proxy : Include distribution specific tasks] ***************\nFriday 13 March 2026  00:28:33 +0000 (0:00:02.859)       0:00:10.211 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.proxy : Configure proxy parameters for apt] ****************\nFriday 13 March 2026  00:28:33 +0000 (0:00:00.236)       0:00:10.447 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.proxy : Set system wide settings in environment file] ******\nFriday 13 March 2026  00:28:34 +0000 (0:00:01.106)       0:00:11.554 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.proxy : Remove system wide settings in environment file] ***\nFriday 13 March 2026  00:28:35 +0000 (0:00:00.587)       0:00:12.141 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nFriday 13 March 2026  00:28:36 +0000 (0:00:00.546)       0:00:12.687 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nFriday 13 March 2026  00:28:36 +0000 (0:00:00.215)       0:00:12.903 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nFriday 13 March 2026  00:28:36 +0000 (0:00:00.253)       0:00:13.157 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nFriday 13 March 2026  00:28:36 +0000 (0:00:00.356)       0:00:13.513 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nFriday 13 March 2026  00:28:38 +0000 (0:00:01.467)       0:00:14.981 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nFriday 13 March 2026  00:28:38 +0000 (0:00:00.227)       0:00:15.209 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nFriday 13 March 2026  00:28:39 +0000 (0:00:00.516)       0:00:15.726 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nFriday 13 March 2026  00:28:39 +0000 (0:00:00.235)       0:00:15.961 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nFriday 13 March 2026  00:28:39 +0000 (0:00:00.592)       0:00:16.553 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nFriday 13 March 2026  00:28:41 +0000 (0:00:01.147)       0:00:17.701 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nFriday 13 March 2026  00:28:42 +0000 (0:00:01.047)       0:00:18.748 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nFriday 13 March 2026  00:28:42 +0000 (0:00:00.301)       0:00:19.050 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nFriday 13 March 2026  00:28:43 +0000 (0:00:01.417)       0:00:20.467 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nFriday 13 March 2026  00:28:44 +0000 (0:00:00.208)       0:00:20.676 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nFriday 13 March 2026  00:28:44 +0000 (0:00:00.217)       0:00:20.893 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nFriday 13 March 2026  00:28:44 +0000 (0:00:00.219)       0:00:21.113 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nFriday 13 March 2026  00:28:44 +0000 (0:00:00.266)       0:00:21.379 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nFriday 13 March 2026  00:28:45 +0000 (0:00:00.553)       0:00:21.933 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nFriday 13 March 2026  00:28:45 +0000 (0:00:00.226)       0:00:22.160 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nFriday 13 March 2026  00:28:46 +0000 (0:00:01.022)       0:00:23.182 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nFriday 13 March 2026  00:28:47 +0000 (0:00:00.618)       0:00:23.800 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nFriday 13 March 2026  00:28:48 +0000 (0:00:01.138)       0:00:24.939 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Gather variables for each operating system] *****\nFriday 13 March 2026  00:29:06 +0000 (0:00:18.642)       0:00:43.581 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****\nFriday 13 March 2026  00:29:07 +0000 (0:00:00.249)       0:00:43.831 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***\nFriday 13 March 2026  00:29:07 +0000 (0:00:00.203)       0:00:44.035 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.rsyslog : Include distribution specific install tasks] ****\nFriday 13 March 2026  00:29:07 +0000 (0:00:00.201)       0:00:44.237 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.rsyslog : Install rsyslog package] ************************\nFriday 13 March 2026  00:29:07 +0000 (0:00:00.281)       0:00:44.518 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********\nFriday 13 March 2026  00:29:09 +0000 (0:00:01.825)       0:00:46.343 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Manage rsyslog service] *************************\nFriday 13 March 2026  00:29:10 +0000 (0:00:00.994)       0:00:47.338 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include fluentd tasks] **************************\nFriday 13 March 2026  00:29:11 +0000 (0:00:00.848)       0:00:48.186 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***\nFriday 13 March 2026  00:29:11 +0000 (0:00:00.290)       0:00:48.477 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.rsyslog : Include additional log server tasks] ************\nFriday 13 March 2026  00:29:12 +0000 (0:00:01.144)       0:00:49.621 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.rsyslog : Include logrotate tasks] ************************\nFriday 13 March 2026  00:29:13 +0000 (0:00:00.212)       0:00:49.834 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.rsyslog : Ensure logrotate package is installed] **********\nFriday 13 March 2026  00:29:13 +0000 (0:00:00.261)       0:00:50.096 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************\nFriday 13 March 2026  00:29:15 +0000 (0:00:01.782)       0:00:51.878 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.systohc : Install util-linux-extra package] ****************\nFriday 13 March 2026  00:29:16 +0000 (0:00:01.052)       0:00:52.930 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.systohc : Sync hardware clock] *****************************\nFriday 13 March 2026  00:29:27 +0000 (0:00:11.359)       0:01:04.289 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.configfs : Start sys-kernel-config mount] ******************\nFriday 13 March 2026  00:29:28 +0000 (0:00:00.996)       0:01:05.286 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Gather variables for each operating system] *****\nFriday 13 March 2026  00:29:29 +0000 (0:00:00.896)       0:01:06.183 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***\nFriday 13 March 2026  00:29:29 +0000 (0:00:00.197)       0:01:06.381 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.packages : Include distribution specific package tasks] ****\nFriday 13 March 2026  00:29:29 +0000 (0:00:00.206)       0:01:06.587 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.packages : Install needrestart package] ********************\nFriday 13 March 2026  00:29:30 +0000 (0:00:00.274)       0:01:06.862 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.packages : Set needrestart mode] ***************************\nFriday 13 March 2026  00:29:32 +0000 (0:00:01.926)       0:01:08.788 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***\nFriday 13 March 2026  00:29:32 +0000 (0:00:00.550)       0:01:09.339 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.packages : Update package cache] ***************************\nFriday 13 March 2026  00:29:32 +0000 (0:00:00.190)       0:01:09.529 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Download upgrade packages] **********************\nFriday 13 March 2026  00:29:34 +0000 (0:00:01.272)       0:01:10.802 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Upgrade packages] *******************************\nFriday 13 March 2026  00:29:36 +0000 (0:00:01.943)       0:01:12.746 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.packages : Download required packages] *********************\nFriday 13 March 2026  00:29:39 +0000 (0:00:03.217)       0:01:15.963 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.packages : Install required packages] **********************\nFriday 13 March 2026  00:30:16 +0000 (0:00:37.340)       0:01:53.304 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.packages : Remove useless packages from the cache] *********\nFriday 13 March 2026  00:31:38 +0000 (0:01:21.612)       0:03:14.917 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.packages : Remove dependencies that are no longer required] ***\nFriday 13 March 2026  00:31:40 +0000 (0:00:01.758)       0:03:16.675 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sysctl : Include sysctl tasks] *****************************\nFriday 13 March 2026  00:31:49 +0000 (0:00:09.951)       0:03:26.627 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********\nFriday 13 March 2026  00:31:50 +0000 (0:00:00.424)       0:03:27.051 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************\nFriday 13 March 2026  00:31:51 +0000 (0:00:00.661)       0:03:27.712 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************\nFriday 13 March 2026  00:31:56 +0000 (0:00:05.914)       0:03:33.627 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************\nFriday 13 March 2026  00:31:58 +0000 (0:00:01.492)       0:03:35.119 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on network] *****************\nFriday 13 March 2026  00:31:59 +0000 (0:00:01.447)       0:03:36.567 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************\nFriday 13 March 2026  00:32:01 +0000 (0:00:01.533)       0:03:38.100 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\nTASK [osism.commons.limits : Include limits tasks] *****************************\nFriday 13 March 2026  00:32:02 +0000 (0:00:00.551)       0:03:38.652 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.services : Populate service facts] *************************\nFriday 13 March 2026  00:32:02 +0000 (0:00:00.280)       0:03:38.932 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.services : Check services] *********************************\nFriday 13 March 2026  00:32:07 +0000 (0:00:05.359)       0:03:44.292 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.services : Start/enable required services] *****************\nFriday 13 March 2026  00:32:07 +0000 (0:00:00.302)       0:03:44.595 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m\n\nTASK [osism.commons.motd : Include distribution specific configure tasks] ******\nFriday 13 March 2026  00:32:09 +0000 (0:00:01.119)       0:03:45.714 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.motd : Remove update-motd package] *************************\nFriday 13 March 2026  00:32:09 +0000 (0:00:00.387)       0:03:46.102 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************\nFriday 13 March 2026  00:32:10 +0000 (0:00:01.276)       0:03:47.378 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Disable the dynamic motd-news service] **************\nFriday 13 March 2026  00:32:11 +0000 (0:00:00.645)       0:03:48.024 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********\nFriday 13 March 2026  00:32:11 +0000 (0:00:00.615)       0:03:48.639 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.motd : Remove pam_motd.so rule] ****************************\nFriday 13 March 2026  00:32:12 +0000 (0:00:00.554)       0:03:49.194 ********** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360330.6471221, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360347.723246, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360324.922893, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360319.1326118, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360323.0768747, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360353.9467874, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360335.4581244, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [osism.commons.motd : Copy motd file] *************************************\nFriday 13 March 2026  00:32:13 +0000 (0:00:01.015)       0:03:50.209 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue file] ************************************\nFriday 13 March 2026  00:32:14 +0000 (0:00:01.148)       0:03:51.358 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue.net file] ********************************\nFriday 13 March 2026  00:32:15 +0000 (0:00:01.117)       0:03:52.476 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to print the motd] ********************\nFriday 13 March 2026  00:32:16 +0000 (0:00:01.030)       0:03:53.506 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to not print the motd] ****************\nFriday 13 March 2026  00:32:17 +0000 (0:00:00.248)       0:03:53.755 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.rng : Include distribution specific install tasks] ********\nFriday 13 March 2026  00:32:17 +0000 (0:00:00.691)       0:03:54.447 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.rng : Install rng package] ********************************\nFriday 13 March 2026  00:32:18 +0000 (0:00:00.378)       0:03:54.826 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.rng : Remove haveged package] *****************************\nFriday 13 March 2026  00:32:26 +0000 (0:00:08.632)       0:04:03.458 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rng : Manage rng service] *********************************\nFriday 13 March 2026  00:32:28 +0000 (0:00:01.311)       0:04:04.769 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Gather variables for each operating system] ******\nFriday 13 March 2026  00:32:29 +0000 (0:00:01.104)       0:04:05.874 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***\nFriday 13 March 2026  00:32:29 +0000 (0:00:00.276)       0:04:06.150 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***\nFriday 13 March 2026  00:32:29 +0000 (0:00:00.291)       0:04:06.442 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Populate service facts] **************************\nFriday 13 March 2026  00:32:30 +0000 (0:00:00.307)       0:04:06.749 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Include distribution specific timer tasks] *******\nFriday 13 March 2026  00:32:34 +0000 (0:00:04.610)       0:04:11.360 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.cleanup : Disable apt-daily timers] ************************\nFriday 13 March 2026  00:32:35 +0000 (0:00:00.398)       0:04:11.759 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Include service tasks] ***************************\nFriday 13 March 2026  00:32:35 +0000 (0:00:00.345)       0:04:12.105 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup services] ********************************\nFriday 13 March 2026  00:32:35 +0000 (0:00:00.442)       0:04:12.547 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.cleanup : Include packages tasks] **************************\nFriday 13 March 2026  00:32:36 +0000 (0:00:00.339)       0:04:12.887 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup installed packages] **********************\nFriday 13 March 2026  00:32:36 +0000 (0:00:00.382)       0:04:13.269 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloudinit package] ************************\nFriday 13 March 2026  00:33:09 +0000 (0:00:32.796)       0:04:46.066 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********\nFriday 13 March 2026  00:33:18 +0000 (0:00:08.631)       0:04:54.697 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove useless packages from the cache] **********\nFriday 13 March 2026  00:33:26 +0000 (0:00:08.069)       0:05:02.767 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***\nFriday 13 March 2026  00:33:27 +0000 (0:00:01.577)       0:05:04.344 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.cleanup : Include cloudinit tasks] *************************\nFriday 13 March 2026  00:33:33 +0000 (0:00:05.897)       0:05:10.242 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******\nFriday 13 March 2026  00:33:33 +0000 (0:00:00.373)       0:05:10.615 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Install tzdata package] *************************\nFriday 13 March 2026  00:33:34 +0000 (0:00:00.624)       0:05:11.239 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.timezone : Set timezone to UTC] ****************************\nFriday 13 March 2026  00:33:36 +0000 (0:00:01.774)       0:05:13.014 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Create /etc/adjtime file] ***********************\nFriday 13 March 2026  00:33:37 +0000 (0:00:00.669)       0:05:13.684 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************\nFriday 13 March 2026  00:33:37 +0000 (0:00:00.209)       0:05:13.893 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Gather variables for each operating system] ******\nFriday 13 March 2026  00:33:37 +0000 (0:00:00.300)       0:05:14.194 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Set docker_version variable to default value] ****\nFriday 13 March 2026  00:33:37 +0000 (0:00:00.261)       0:05:14.455 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Set docker_cli_version variable to default value] ***\nFriday 13 March 2026  00:33:38 +0000 (0:00:00.228)       0:05:14.684 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Print used docker version] ***********************\nFriday 13 March 2026  00:33:38 +0000 (0:00:00.264)       0:05:14.948 ********** \n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Print used docker cli version] *******************\nFriday 13 March 2026  00:33:38 +0000 (0:00:00.227)       0:05:15.175 ********** \n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Include block storage tasks] *********************\nFriday 13 March 2026  00:33:38 +0000 (0:00:00.234)       0:05:15.409 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Include zram storage tasks] **********************\nFriday 13 March 2026  00:33:38 +0000 (0:00:00.227)       0:05:15.637 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Include docker install tasks] ********************\nFriday 13 March 2026  00:33:39 +0000 (0:00:00.224)       0:05:15.861 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.docker : Remove old architecture-dependent repository] ****\nFriday 13 March 2026  00:33:39 +0000 (0:00:00.443)       0:05:16.305 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Gather package facts] ****************************\nFriday 13 March 2026  00:33:40 +0000 (0:00:00.774)       0:05:17.079 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.docker : Check whether packages are installed that should not be installed] ***\nFriday 13 March 2026  00:33:43 +0000 (0:00:02.938)       0:05:20.017 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install apt-transport-https package] *************\nFriday 13 March 2026  00:33:43 +0000 (0:00:00.573)       0:05:20.591 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Add repository gpg key] **************************\nFriday 13 March 2026  00:33:50 +0000 (0:00:06.564)       0:05:27.155 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Add repository] **********************************\nFriday 13 March 2026  00:33:51 +0000 (0:00:01.010)       0:05:28.166 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Update package cache] ****************************\nFriday 13 March 2026  00:33:59 +0000 (0:00:08.197)       0:05:36.364 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Pin docker package version] **********************\nFriday 13 March 2026  00:34:02 +0000 (0:00:03.185)       0:05:39.549 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Pin docker-cli package version] ******************\nFriday 13 March 2026  00:34:04 +0000 (0:00:01.200)       0:05:40.750 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Unlock containerd package] ***********************\nFriday 13 March 2026  00:34:05 +0000 (0:00:01.452)       0:05:42.202 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install containerd package] **********************\nFriday 13 March 2026  00:34:06 +0000 (0:00:01.004)       0:05:43.207 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Lock containerd package] *************************\nFriday 13 March 2026  00:34:16 +0000 (0:00:09.642)       0:05:52.849 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install docker-cli package] **********************\nFriday 13 March 2026  00:34:17 +0000 (0:00:00.952)       0:05:53.802 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Install docker package] **************************\nFriday 13 March 2026  00:34:26 +0000 (0:00:09.176)       0:06:02.979 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Unblock installation of python docker packages] ***\nFriday 13 March 2026  00:34:37 +0000 (0:00:10.797)       0:06:13.777 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package] ******************\nFriday 13 March 2026  00:34:38 +0000 (0:00:01.224)       0:06:15.001 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package from Debian Sid] ***\nFriday 13 March 2026  00:34:38 +0000 (0:00:00.529)       0:06:15.531 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***\nFriday 13 March 2026  00:34:42 +0000 (0:00:03.866)       0:06:19.397 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***\nFriday 13 March 2026  00:34:43 +0000 (0:00:00.671)       0:06:20.069 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***\nFriday 13 March 2026  00:34:43 +0000 (0:00:00.563)       0:06:20.632 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***\nFriday 13 March 2026  00:34:44 +0000 (0:00:00.510)       0:06:21.143 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install packages required by docker login] *******\nFriday 13 March 2026  00:34:45 +0000 (0:00:00.513)       0:06:21.656 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Ensure that some packages are not installed] *****\nFriday 13 March 2026  00:34:45 +0000 (0:00:00.553)       0:06:22.209 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Include config tasks] ****************************\nFriday 13 March 2026  00:34:47 +0000 (0:00:01.825)       0:06:24.035 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.docker : Create plugins directory] ************************\nFriday 13 March 2026  00:34:48 +0000 (0:00:00.886)       0:06:24.922 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Create systemd overlay directory] ****************\nFriday 13 March 2026  00:34:49 +0000 (0:00:00.842)       0:06:25.764 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Copy systemd overlay file] ***********************\nFriday 13 March 2026  00:34:49 +0000 (0:00:00.865)       0:06:26.629 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***\nFriday 13 March 2026  00:34:51 +0000 (0:00:01.532)       0:06:28.161 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy limits configuration file] ******************\nFriday 13 March 2026  00:34:52 +0000 (0:00:01.216)       0:06:29.378 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Copy daemon.json configuration file] *************\nFriday 13 March 2026  00:34:54 +0000 (0:00:01.270)       0:06:30.649 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Include service tasks] ***************************\nFriday 13 March 2026  00:34:55 +0000 (0:00:01.350)       0:06:31.999 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon] ***************************\nFriday 13 March 2026  00:34:56 +0000 (0:00:00.971)       0:06:32.970 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Manage service] **********************************\nFriday 13 March 2026  00:34:57 +0000 (0:00:01.382)       0:06:34.353 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Manage docker socket service] ********************\nFriday 13 March 2026  00:34:58 +0000 (0:00:01.135)       0:06:35.489 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Manage containerd service] ***********************\nFriday 13 March 2026  00:34:59 +0000 (0:00:01.116)       0:06:36.606 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Include bootstrap tasks] *************************\nFriday 13 March 2026  00:35:01 +0000 (0:00:01.262)       0:06:37.868 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.843)       0:06:38.711 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.037)       0:06:38.749 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.043)       0:06:38.793 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.039)       0:06:38.832 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.037)       0:06:38.870 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.044)       0:06:38.915 ********** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.046)       0:06:38.961 ********** \n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nFriday 13 March 2026  00:35:02 +0000 (0:00:00.038)       0:06:38.999 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************\nFriday 13 March 2026  00:35:03 +0000 (0:00:01.159)       0:06:40.158 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********\nFriday 13 March 2026  00:35:04 +0000 (0:00:01.459)       0:06:41.618 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Restart docker service] ***************\nFriday 13 March 2026  00:35:06 +0000 (0:00:01.306)       0:06:42.924 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****\nFriday 13 March 2026  00:35:08 +0000 (0:00:02.178)       0:06:45.103 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Add user to docker group] ************************\nFriday 13 March 2026  00:35:08 +0000 (0:00:00.081)       0:06:45.184 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Log into private registry and force re-authorization] ***\nFriday 13 March 2026  00:35:09 +0000 (0:00:00.928)       0:06:46.113 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Include facts tasks] *****************************\nFriday 13 March 2026  00:35:09 +0000 (0:00:00.522)       0:06:46.635 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.docker : Create facts directory] **************************\nFriday 13 March 2026  00:35:11 +0000 (0:00:01.111)       0:06:47.747 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Copy docker fact files] **************************\nFriday 13 March 2026  00:35:11 +0000 (0:00:00.796)       0:06:48.544 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m\n\nTASK [osism.commons.docker_compose : This install type is not supported] *******\nFriday 13 March 2026  00:35:14 +0000 (0:00:02.451)       0:06:50.995 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.docker_compose : Include distribution specific install tasks] ***\nFriday 13 March 2026  00:35:14 +0000 (0:00:00.520)       0:06:51.516 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***\nFriday 13 March 2026  00:35:15 +0000 (0:00:00.762)       0:06:52.279 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******\nFriday 13 March 2026  00:35:16 +0000 (0:00:00.815)       0:06:53.094 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose binary] *************\nFriday 13 March 2026  00:35:17 +0000 (0:00:00.985)       0:06:54.080 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.docker_compose : Uninstall docker-compose package] *********\nFriday 13 March 2026  00:35:17 +0000 (0:00:00.493)       0:06:54.573 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose script] ***************\nFriday 13 March 2026  00:35:19 +0000 (0:00:01.483)       0:06:56.056 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****\nFriday 13 March 2026  00:35:19 +0000 (0:00:00.486)       0:06:56.543 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********\nFriday 13 March 2026  00:35:28 +0000 (0:00:08.455)       0:07:04.998 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Enable osism.target] **********************\nFriday 13 March 2026  00:35:29 +0000 (0:00:01.438)       0:07:06.436 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****\nFriday 13 March 2026  00:35:31 +0000 (0:00:01.559)       0:07:07.996 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nFriday 13 March 2026  00:35:32 +0000 (0:00:01.508)       0:07:09.504 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nFriday 13 March 2026  00:35:33 +0000 (0:00:00.837)       0:07:10.341 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Check minimum and maximum number of servers] *****\nFriday 13 March 2026  00:35:34 +0000 (0:00:00.822)       0:07:11.164 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Gather variables for each operating system] ******\nFriday 13 March 2026  00:35:35 +0000 (0:00:00.539)       0:07:11.703 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***\nFriday 13 March 2026  00:35:35 +0000 (0:00:00.551)       0:07:12.255 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_key_file variable to default value] ***\nFriday 13 March 2026  00:35:36 +0000 (0:00:00.756)       0:07:13.012 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Populate service facts] **************************\nFriday 13 March 2026  00:35:36 +0000 (0:00:00.513)       0:07:13.526 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.chrony : Manage timesyncd service] ************************\nFriday 13 March 2026  00:35:42 +0000 (0:00:05.378)       0:07:18.904 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Include distribution specific install tasks] *****\nFriday 13 March 2026  00:35:42 +0000 (0:00:00.481)       0:07:19.386 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.chrony : Install package] *********************************\nFriday 13 March 2026  00:35:43 +0000 (0:00:00.800)       0:07:20.187 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Manage chrony service] ***************************\nFriday 13 March 2026  00:35:45 +0000 (0:00:01.766)       0:07:21.954 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Check if configuration file exists] **************\nFriday 13 March 2026  00:35:47 +0000 (0:00:02.035)       0:07:23.989 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.chrony : Copy configuration file] *************************\nFriday 13 March 2026  00:35:48 +0000 (0:00:00.811)       0:07:24.801 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\nTASK [osism.services.lldpd : Include distribution specific install tasks] ******\nFriday 13 March 2026  00:35:49 +0000 (0:00:01.814)       0:07:26.616 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.lldpd : Install lldpd package] ****************************\nFriday 13 March 2026  00:35:50 +0000 (0:00:00.789)       0:07:27.405 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.lldpd : Manage lldpd service] *****************************\nFriday 13 March 2026  00:36:00 +0000 (0:00:09.257)       0:07:36.662 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********\nFriday 13 March 2026  00:36:02 +0000 (0:00:01.985)       0:07:38.648 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************\nFriday 13 March 2026  00:36:03 +0000 (0:00:01.257)       0:07:39.906 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply bootstrap role part 2] *********************************************\n\nTASK [Include hardening role] **************************************************\nFriday 13 March 2026  00:36:04 +0000 (0:00:01.177)       0:07:41.084 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Apply bootstrap roles part 3] ********************************************\n\nTASK [osism.services.journald : Copy configuration file] ***********************\nFriday 13 March 2026  00:36:05 +0000 (0:00:00.722)       0:07:41.806 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.journald : Manage journald service] ***********************\nFriday 13 March 2026  00:36:06 +0000 (0:00:01.413)       0:07:43.220 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Include auditd role] *****************************************************\nFriday 13 March 2026  00:36:07 +0000 (0:00:01.344)       0:07:44.564 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [Include smartd role] *****************************************************\nFriday 13 March 2026  00:36:08 +0000 (0:00:00.660)       0:07:45.224 ********** \n\u001b[0;36mincluded: osism.services.smartd for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.smartd : Include distribution specific install tasks] *****\nFriday 13 March 2026  00:36:09 +0000 (0:00:00.835)       0:07:46.060 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.services.smartd : Install smartmontools package] *******************\nFriday 13 March 2026  00:36:10 +0000 (0:00:00.786)       0:07:46.846 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Create /var/log/smartd directory] ****************\nFriday 13 March 2026  00:36:19 +0000 (0:00:09.117)       0:07:55.963 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.smartd : Copy smartmontools configuration file] ***********\nFriday 13 March 2026  00:36:20 +0000 (0:00:00.711)       0:07:56.674 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Manage smartd service] ***************************\nFriday 13 March 2026  00:36:21 +0000 (0:00:01.219)       0:07:57.894 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.journald : Restart journald service] ***********\nFriday 13 March 2026  00:36:22 +0000 (0:00:01.715)       0:07:59.609 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************\nFriday 13 March 2026  00:36:24 +0000 (0:00:01.266)       0:08:00.876 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Set state bootstrap] *****************************************************\n\nTASK [Set osism.bootstrap.status fact] *****************************************\nFriday 13 March 2026  00:36:25 +0000 (0:00:01.033)       0:08:01.909 ********** \n\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nFriday 13 March 2026  00:36:26 +0000 (0:00:00.911)       0:08:02.821 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nFriday 13 March 2026  00:36:26 +0000 (0:00:00.777)       0:08:03.599 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set osism.bootstrap.timestamp fact] **************************************\nFriday 13 March 2026  00:36:28 +0000 (0:00:01.134)       0:08:04.733 ********** \n\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nFriday 13 March 2026  00:36:29 +0000 (0:00:00.922)       0:08:05.656 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nFriday 13 March 2026  00:36:29 +0000 (0:00:00.829)       0:08:06.485 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:36:30 +0000 (0:00:01.039)       0:08:07.525 ********** \n=============================================================================== \nosism.commons.packages : Install required packages --------------------- 81.61s\nosism.commons.packages : Download required packages -------------------- 37.34s\nosism.commons.cleanup : Cleanup installed packages --------------------- 32.80s\nosism.commons.repository : Update package cache ------------------------ 18.64s\nosism.commons.systohc : Install util-linux-extra package --------------- 11.36s\nosism.services.docker : Install docker package ------------------------- 10.80s\nosism.commons.packages : Remove dependencies that are no longer required --- 9.95s\nosism.services.docker : Install containerd package ---------------------- 9.64s\nosism.services.lldpd : Install lldpd package ---------------------------- 9.26s\nosism.services.docker : Install docker-cli package ---------------------- 9.18s\nosism.services.smartd : Install smartmontools package ------------------- 9.12s\nosism.services.rng : Install rng package -------------------------------- 8.63s\nosism.commons.cleanup : Remove cloudinit package ------------------------ 8.63s\nosism.commons.docker_compose : Install docker-compose-plugin package ---- 8.46s\nosism.services.docker : Add repository ---------------------------------- 8.20s\nosism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.07s\nosism.services.docker : Install apt-transport-https package ------------- 6.56s\nosism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 5.91s\nosism.commons.cleanup : Remove dependencies that are no longer required --- 5.90s\nosism.services.chrony : Populate service facts -------------------------- 5.38s\n+ osism apply fail2ban\n\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m\n\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dd2ed9d3-1d42-40d1-934e-9f5944ffab6b (fail2ban) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dd2ed9d3-1d42-40d1-934e-9f5944ffab6b (fail2ban) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role fail2ban] *****************************************************\n\nTASK [osism.services.fail2ban : Include distribution specific install tasks] ***\nFriday 13 March 2026  00:36:48 +0000 (0:00:00.242)       0:00:00.242 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.fail2ban : Install fail2ban package] **********************\nFriday 13 March 2026  00:36:49 +0000 (0:00:01.011)       0:00:01.253 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.fail2ban : Copy configuration files] **********************\nFriday 13 March 2026  00:36:59 +0000 (0:00:10.648)       0:00:11.902 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Manage fail2ban service] ***********************\nFriday 13 March 2026  00:37:01 +0000 (0:00:01.341)       0:00:13.243 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Reload fail2ban configuration] *****************\nFriday 13 March 2026  00:37:02 +0000 (0:00:01.335)       0:00:14.578 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:37:03 +0000 (0:00:01.442)       0:00:16.021 ********** \n=============================================================================== \nosism.services.fail2ban : Install fail2ban package --------------------- 10.65s\nosism.services.fail2ban : Reload fail2ban configuration ----------------- 1.44s\nosism.services.fail2ban : Copy configuration files ---------------------- 1.34s\nosism.services.fail2ban : Manage fail2ban service ----------------------- 1.33s\nosism.services.fail2ban : Include distribution specific install tasks --- 1.01s\n+ [[ -e /etc/redhat-release ]]\n+ osism apply network\n\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m\n\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 43595a2f-1241-4c1c-91aa-86c7a57b59e5 (network) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 43595a2f-1241-4c1c-91aa-86c7a57b59e5 (network) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role network] ******************************************************\n\nTASK [osism.commons.network : Gather variables for each operating system] ******\nFriday 13 March 2026  00:37:19 +0000 (0:00:00.187)       0:00:00.187 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include type specific tasks] *********************\nFriday 13 March 2026  00:37:20 +0000 (0:00:00.611)       0:00:00.799 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Install required packages] ***********************\nFriday 13 March 2026  00:37:21 +0000 (0:00:01.070)       0:00:01.870 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.network : Remove ifupdown package] *************************\nFriday 13 March 2026  00:37:23 +0000 (0:00:01.958)       0:00:03.829 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.network : Create required directories] *********************\nFriday 13 March 2026  00:37:25 +0000 (0:00:01.732)       0:00:05.561 ********** \n\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m\n\nTASK [osism.commons.network : Prepare netplan configuration template] **********\nFriday 13 March 2026  00:37:26 +0000 (0:00:00.915)       0:00:06.477 ********** \n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Copy netplan configuration] **********************\nFriday 13 March 2026  00:37:29 +0000 (0:00:03.080)       0:00:09.557 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove netplan configuration template] ***********\nFriday 13 March 2026  00:37:30 +0000 (0:00:01.596)       0:00:11.154 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Check if path for interface file exists] *********\nFriday 13 March 2026  00:37:32 +0000 (0:00:01.649)       0:00:12.803 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy interfaces file] ****************************\nFriday 13 March 2026  00:37:33 +0000 (0:00:01.065)       0:00:13.868 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Install package networkd-dispatcher] *************\nFriday 13 March 2026  00:37:34 +0000 (0:00:00.637)       0:00:14.506 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.network : Copy dispatcher scripts] *************************\nFriday 13 March 2026  00:37:36 +0000 (0:00:02.258)       0:00:16.764 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m\n\nTASK [osism.commons.network : Manage service networkd-dispatcher] **************\nFriday 13 March 2026  00:37:37 +0000 (0:00:00.761)       0:00:17.526 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include cleanup tasks] ***************************\nFriday 13 March 2026  00:37:38 +0000 (0:00:01.505)       0:00:19.032 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nFriday 13 March 2026  00:37:39 +0000 (0:00:01.068)       0:00:20.100 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Set network_configured_files fact] ***************\nFriday 13 March 2026  00:37:40 +0000 (0:00:01.014)       0:00:21.114 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nFriday 13 March 2026  00:37:41 +0000 (0:00:00.567)       0:00:21.681 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\nTASK [osism.commons.network : Include dummy interfaces] ************************\nFriday 13 March 2026  00:37:42 +0000 (0:00:01.075)       0:00:22.757 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include vxlan interfaces] ************************\nFriday 13 March 2026  00:37:42 +0000 (0:00:00.660)       0:00:23.417 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-manager, testbed-node-1, testbed-node-0, testbed-node-4, testbed-node-2, testbed-node-3, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd netdev files] ************\nFriday 13 March 2026  00:37:47 +0000 (0:00:04.420)       0:00:27.838 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd network files] ***********\nFriday 13 March 2026  00:37:52 +0000 (0:00:04.944)       0:00:32.783 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Include networkd cleanup tasks] ******************\nFriday 13 March 2026  00:37:57 +0000 (0:00:05.025)       0:00:37.808 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nFriday 13 March 2026  00:37:58 +0000 (0:00:00.988)       0:00:38.797 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nFriday 13 March 2026  00:37:59 +0000 (0:00:01.042)       0:00:39.840 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include network extra init] **********************\nFriday 13 March 2026  00:38:00 +0000 (0:00:00.771)       0:00:40.611 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init script] ****************\nFriday 13 March 2026  00:38:01 +0000 (0:00:01.086)       0:00:41.697 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init systemd service] *******\nFriday 13 March 2026  00:38:01 +0000 (0:00:00.545)       0:00:42.242 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Enable and start network-extra-init service] *****\nFriday 13 March 2026  00:38:02 +0000 (0:00:00.650)       0:00:42.893 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Disable and stop network-extra-init service] *****\nFriday 13 March 2026  00:38:03 +0000 (0:00:00.582)       0:00:43.476 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init systemd service] *******\nFriday 13 March 2026  00:38:04 +0000 (0:00:01.597)       0:00:45.074 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init script] ****************\nFriday 13 March 2026  00:38:05 +0000 (0:00:00.993)       0:00:46.067 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************\nFriday 13 March 2026  00:38:07 +0000 (0:00:02.363)       0:00:48.431 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********\nFriday 13 March 2026  00:38:08 +0000 (0:00:00.803)       0:00:49.234 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:38:09 +0000 (0:00:00.521)       0:00:49.756 ********** \n=============================================================================== \nosism.commons.network : Create systemd networkd network files ----------- 5.03s\nosism.commons.network : Create systemd networkd netdev files ------------ 4.94s\nosism.commons.network : Include vxlan interfaces ------------------------ 4.42s\nosism.commons.network : Prepare netplan configuration template ---------- 3.08s\nosism.commons.network : Remove network-extra-init script ---------------- 2.36s\nosism.commons.network : Install package networkd-dispatcher ------------- 2.26s\nosism.commons.network : Install required packages ----------------------- 1.96s\nosism.commons.network : Remove ifupdown package ------------------------- 1.73s\nosism.commons.network : Remove netplan configuration template ----------- 1.65s\nosism.commons.network : Disable and stop network-extra-init service ----- 1.60s\nosism.commons.network : Copy netplan configuration ---------------------- 1.60s\nosism.commons.network : Manage service networkd-dispatcher -------------- 1.51s\nosism.commons.network : Include network extra init ---------------------- 1.09s\nosism.commons.network : Remove unused configuration files --------------- 1.08s\nosism.commons.network : Include type specific tasks --------------------- 1.07s\nosism.commons.network : Include cleanup tasks --------------------------- 1.07s\nosism.commons.network : Check if path for interface file exists --------- 1.07s\nosism.commons.network : List existing configuration files --------------- 1.04s\nosism.commons.network : List existing configuration files --------------- 1.01s\nosism.commons.network : Remove network-extra-init systemd service ------- 0.99s\n+ osism apply wireguard\n\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m\n\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bf55688f-ea1a-405e-95fa-e6c0ceb11bb2 (wireguard) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bf55688f-ea1a-405e-95fa-e6c0ceb11bb2 (wireguard) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role wireguard] ****************************************************\n\nTASK [osism.services.wireguard : Install iptables package] *********************\nFriday 13 March 2026  00:38:25 +0000 (0:00:00.168)       0:00:00.168 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Install wireguard package] ********************\nFriday 13 March 2026  00:38:27 +0000 (0:00:01.144)       0:00:01.312 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create public and private key - server] *******\nFriday 13 March 2026  00:38:32 +0000 (0:00:05.011)       0:00:06.324 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create preshared key] *************************\nFriday 13 March 2026  00:38:32 +0000 (0:00:00.498)       0:00:06.823 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get preshared key] ****************************\nFriday 13 March 2026  00:38:32 +0000 (0:00:00.375)       0:00:07.199 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get public key - server] **********************\nFriday 13 March 2026  00:38:33 +0000 (0:00:00.558)       0:00:07.758 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get private key - server] *********************\nFriday 13 March 2026  00:38:33 +0000 (0:00:00.383)       0:00:08.142 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy wg0.conf configuration file] *************\nFriday 13 March 2026  00:38:34 +0000 (0:00:00.378)       0:00:08.520 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy client configuration files] **************\nFriday 13 March 2026  00:38:35 +0000 (0:00:01.133)       0:00:09.654 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********\nFriday 13 March 2026  00:38:36 +0000 (0:00:00.893)       0:00:10.547 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************\nFriday 13 March 2026  00:38:37 +0000 (0:00:01.661)       0:00:12.209 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:38:38 +0000 (0:00:00.889)       0:00:13.098 ********** \n=============================================================================== \nosism.services.wireguard : Install wireguard package -------------------- 5.01s\nosism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.66s\nosism.services.wireguard : Install iptables package --------------------- 1.14s\nosism.services.wireguard : Copy wg0.conf configuration file ------------- 1.13s\nosism.services.wireguard : Copy client configuration files -------------- 0.89s\nosism.services.wireguard : Restart wg0 service -------------------------- 0.89s\nosism.services.wireguard : Get preshared key ---------------------------- 0.56s\nosism.services.wireguard : Create public and private key - server ------- 0.50s\nosism.services.wireguard : Get public key - server ---------------------- 0.38s\nosism.services.wireguard : Get private key - server --------------------- 0.38s\nosism.services.wireguard : Create preshared key ------------------------- 0.38s\n+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0\r100    14  100    14    0     0    197      0 --:--:-- --:--:-- --:--:--   200\n+ osism apply --environment custom workarounds\n\u001b[32m2026-03-13 00:38:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m\n\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m\n\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f20912d7-2bba-457a-8259-9c7be1649d3c (workarounds) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f20912d7-2bba-457a-8259-9c7be1649d3c (workarounds) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on virtualization_role] ********************************\nFriday 13 March 2026  00:38:55 +0000 (0:00:00.112)       0:00:00.112 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m\n\nPLAY [Apply netplan configuration on the manager node] *************************\n\nTASK [Apply netplan configuration] *********************************************\nFriday 13 March 2026  00:38:56 +0000 (0:00:00.669)       0:00:00.782 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply netplan configuration on all other nodes] **************************\n\nTASK [Apply netplan configuration] *********************************************\nFriday 13 March 2026  00:38:58 +0000 (0:00:02.086)       0:00:02.868 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Add custom CA certificates to non-manager nodes] *************************\n\nTASK [Copy custom CA certificates] *********************************************\nFriday 13 March 2026  00:39:00 +0000 (0:00:01.847)       0:00:04.716 ********** \n\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\nTASK [Run update-ca-certificates] **********************************************\nFriday 13 March 2026  00:39:01 +0000 (0:00:01.422)       0:00:06.138 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Run update-ca-trust] *****************************************************\nFriday 13 March 2026  00:39:05 +0000 (0:00:03.509)       0:00:09.647 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Add a workaround service] ************************************************\n\nTASK [Copy workarounds.sh scripts] *********************************************\nFriday 13 March 2026  00:39:05 +0000 (0:00:00.610)       0:00:10.258 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Copy workarounds systemd unit file] **************************************\nFriday 13 March 2026  00:39:07 +0000 (0:00:01.790)       0:00:12.049 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reload systemd daemon] ***************************************************\nFriday 13 March 2026  00:39:09 +0000 (0:00:01.590)       0:00:13.639 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Enable workarounds.service (Debian)] *************************************\nFriday 13 March 2026  00:39:10 +0000 (0:00:01.486)       0:00:15.126 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Enable and start workarounds.service (RedHat)] ***************************\nFriday 13 March 2026  00:39:12 +0000 (0:00:01.813)       0:00:16.940 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************\n\nTASK [Install python3-docker] **************************************************\nFriday 13 March 2026  00:39:13 +0000 (0:00:00.649)       0:00:17.589 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:39:15 +0000 (0:00:02.843)       0:00:20.432 ********** \n=============================================================================== \nRun update-ca-certificates ---------------------------------------------- 3.51s\nInstall python3-docker -------------------------------------------------- 2.84s\nApply netplan configuration --------------------------------------------- 2.09s\nApply netplan configuration --------------------------------------------- 1.85s\nEnable workarounds.service (Debian) ------------------------------------- 1.81s\nCopy workarounds.sh scripts --------------------------------------------- 1.79s\nCopy workarounds systemd unit file -------------------------------------- 1.59s\nReload systemd daemon --------------------------------------------------- 1.49s\nCopy custom CA certificates --------------------------------------------- 1.42s\nGroup hosts based on virtualization_role -------------------------------- 0.67s\nEnable and start workarounds.service (RedHat) --------------------------- 0.65s\nRun update-ca-trust ----------------------------------------------------- 0.61s\n+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m\n\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4c7df022-a09d-40d7-84a9-9061b406155a (reboot) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4c7df022-a09d-40d7-84a9-9061b406155a (reboot) has been started and output is visible here.\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:32 +0000 (0:00:00.149)       0:00:00.149 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:32 +0000 (0:00:00.096)       0:00:00.246 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:33 +0000 (0:00:00.898)       0:00:01.144 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:33 +0000 (0:00:00.095)       0:00:01.240 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:33 +0000 (0:00:00.085)       0:00:01.325 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:33 +0000 (0:00:00.630)       0:00:01.955 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:34 +0000 (0:00:00.105)       0:00:02.061 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:34 +0000 (0:00:00.154)       0:00:02.216 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:34 +0000 (0:00:00.667)       0:00:02.884 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:34 +0000 (0:00:00.104)       0:00:02.988 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:35 +0000 (0:00:00.094)       0:00:03.083 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:35 +0000 (0:00:00.653)       0:00:03.736 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:35 +0000 (0:00:00.101)       0:00:03.837 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:35 +0000 (0:00:00.096)       0:00:03.934 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:36 +0000 (0:00:00.685)       0:00:04.619 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nFriday 13 March 2026  00:39:36 +0000 (0:00:00.125)       0:00:04.745 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nFriday 13 March 2026  00:39:36 +0000 (0:00:00.109)       0:00:04.854 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nFriday 13 March 2026  00:39:37 +0000 (0:00:00.714)       0:00:05.569 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:39:37 +0000 (0:00:00.033)       0:00:05.602 ********** \n=============================================================================== \nReboot system - do not wait for the reboot to complete ------------------ 4.25s\nExit playbook, if user did not mean to reboot systems ------------------- 0.64s\nReboot system - wait for the reboot to complete ------------------------- 0.57s\n+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-03-13 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m\n\u001b[32m2026-03-13 00:39:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 899bfc0f-3626-4da4-91f0-7044c9d66a42 (wait-for-connection) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:39:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 899bfc0f-3626-4da4-91f0-7044c9d66a42 (wait-for-connection) has been started and output is visible here.\u001b[0m\n\nPLAY [Wait until remote systems are reachable] *********************************\n\nTASK [Wait until remote system is reachable] ***********************************\nFriday 13 March 2026  00:39:53 +0000 (0:00:00.243)       0:00:00.243 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:40:05 +0000 (0:00:11.488)       0:00:11.732 ********** \n=============================================================================== \nWait until remote system is reachable ---------------------------------- 11.49s\n+ osism apply hddtemp\n\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m\n\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca56fc81-55da-4470-a81d-f25cc4eddfbf (hddtemp) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ca56fc81-55da-4470-a81d-f25cc4eddfbf (hddtemp) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role hddtemp] ******************************************************\n\nTASK [osism.services.hddtemp : Gather variables for each operating system] *****\nFriday 13 March 2026  00:40:22 +0000 (0:00:00.233)       0:00:00.233 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific install tasks] ****\nFriday 13 March 2026  00:40:23 +0000 (0:00:00.625)       0:00:00.859 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Remove hddtemp package] *************************\nFriday 13 March 2026  00:40:24 +0000 (0:00:01.005)       0:00:01.864 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************\nFriday 13 March 2026  00:40:25 +0000 (0:00:01.874)       0:00:03.739 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Check if drivetemp module is available] *********\nFriday 13 March 2026  00:40:26 +0000 (0:00:01.027)       0:00:04.766 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************\nFriday 13 March 2026  00:40:28 +0000 (0:00:01.192)       0:00:05.959 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Install lm-sensors] *****************************\nFriday 13 March 2026  00:40:28 +0000 (0:00:00.772)       0:00:06.731 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific service tasks] ****\nFriday 13 March 2026  00:40:42 +0000 (0:00:13.955)       0:00:20.687 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Manage lm-sensors service] **********************\nFriday 13 March 2026  00:40:44 +0000 (0:00:01.118)       0:00:21.806 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:40:45 +0000 (0:00:01.769)       0:00:23.575 ********** \n=============================================================================== \nosism.services.hddtemp : Install lm-sensors ---------------------------- 13.96s\nosism.services.hddtemp : Remove hddtemp package ------------------------- 1.87s\nosism.services.hddtemp : Manage lm-sensors service ---------------------- 1.77s\nosism.services.hddtemp : Check if drivetemp module is available --------- 1.19s\nosism.services.hddtemp : Include distribution specific service tasks ---- 1.12s\nosism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.03s\nosism.services.hddtemp : Include distribution specific install tasks ---- 1.01s\nosism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.77s\nosism.services.hddtemp : Gather variables for each operating system ----- 0.63s\n++ semver latest 7.1.1\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sudo systemctl restart manager.service\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\nARA in ceph-ansible already disabled.\nARA in kolla-ansible already disabled.\nARA in osism-ansible already disabled.\nARA in osism-kubernetes already disabled.\n+ osism apply gather-facts\n\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m\n\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5f6f437b-413b-4f54-9e59-ce9e6a25bddd (gather-facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5f6f437b-413b-4f54-9e59-ce9e6a25bddd (gather-facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:42:21 +0000 (0:00:00.160)       0:00:00.160 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nFriday 13 March 2026  00:42:30 +0000 (0:00:08.252)       0:00:08.412 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:42:30 +0000 (0:00:00.447)       0:00:08.859 ********** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 8.25s\nGather facts for all hosts ---------------------------------------------- 0.45s\n+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack\n+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring\n+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack\n+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring\n+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry\n+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images\n+ [[ false == \\t\\r\\u\\e ]]",
                            "stdout_lines": [
                                "",
                                "# DEPLOY MANAGER",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY MANAGER'",
                                "+ echo",
                                "+ cat /opt/manager-vars.sh",
                                "export NUMBER_OF_NODES=6",
                                "",
                                "export CEPH_VERSION=reef",
                                "export CONFIGURATION_VERSION=main",
                                "export MANAGER_VERSION=latest",
                                "export OPENSTACK_VERSION=2025.1",
                                "",
                                "export ARA=false",
                                "export DEPLOY_MODE=manager",
                                "export TEMPEST=true",
                                "export IS_ZUUL=true",
                                "",
                                "export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "export EXTERNAL_API=false",
                                "",
                                "export IMAGE_USER=ubuntu",
                                "export IMAGE_NODE_USER=ubuntu",
                                "",
                                "export CEPH_STACK=ceph-ansible",
                                "",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver",
                                "+ docker version",
                                "Client: Docker Engine - Community",
                                " Version:           27.5.1",
                                " API version:       1.47",
                                " Go version:        go1.22.11",
                                " Git commit:        9f9e405",
                                " Built:             Wed Jan 22 13:41:48 2025",
                                " OS/Arch:           linux/amd64",
                                " Context:           default",
                                "",
                                "Server: Docker Engine - Community",
                                " Engine:",
                                "  Version:          27.5.1",
                                "  API version:      1.47 (minimum version 1.24)",
                                "  Go version:       go1.22.11",
                                "  Git commit:       4c9b3b0",
                                "  Built:            Wed Jan 22 13:41:48 2025",
                                "  OS/Arch:          linux/amd64",
                                "  Experimental:     false",
                                " containerd:",
                                "  Version:          v2.2.2",
                                "  GitCommit:        301b2dac98f15c27117da5c8af12118a041a31d9",
                                " runc:",
                                "  Version:          1.3.4",
                                "  GitCommit:        v1.3.4-0-gd6d73eb8",
                                " docker-init:",
                                "  Version:          0.19.0",
                                "  GitCommit:        de40ad0",
                                "+ sh -c /opt/configuration/scripts/deploy/000-manager.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-ceph-version.sh reef",
                                "+ set -e",
                                "+ VERSION=reef",
                                "++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n ceph_version: reef ]]",
                                "+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ /opt/configuration/scripts/set-openstack-version.sh 2025.1",
                                "+ set -e",
                                "+ VERSION=2025.1",
                                "++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n openstack_version: 2024.2 ]]",
                                "+ sed -i 's/openstack_version: .*/openstack_version: 2025.1/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ ceph-ansible == \\r\\o\\o\\k ]]",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo 'enable_osism_kubernetes: true'",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 2025.1 2025.1",
                                "+ [[ 0 -ge 0 ]]",
                                "+ sed -i '/^om_enable_rabbitmq_high_availability:/d' /opt/configuration/environments/kolla/configuration.yml",
                                "+ sed -i '/^om_enable_rabbitmq_quorum_queues:/d' /opt/configuration/environments/kolla/configuration.yml",
                                "+ /opt/configuration/scripts/enable-resource-nodes.sh",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml",
                                "",
                                "PLAY [Copy custom facts] *******************************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY [Before the deployment of the manager] ************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get /opt/manager-vars.sh] ************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add ara_server_mariadb_volume_type parameter] ****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Add netbox_enable parameter] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Check if /etc/OTC_region exist] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add nova_compute_virt_type parameter] ************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Apply role traefik] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply traefik role] ******************************************************",
                                "included: osism.services.traefik for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/traefik)",
                                "changed: [testbed-manager] => (item=/opt/traefik/certificates)",
                                "changed: [testbed-manager] => (item=/opt/traefik/configuration)",
                                "",
                                "TASK [osism.services.traefik : Copy configuration files] ***********************",
                                "changed: [testbed-manager] => (item=traefik.yml)",
                                "changed: [testbed-manager] => (item=traefik.env)",
                                "changed: [testbed-manager] => (item=certificates.yml)",
                                "",
                                "TASK [osism.services.traefik : Copy certificate cert files] ********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy certificate key files] *********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy dynamic configuration] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Remove dynamic configuration] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create traefik external network] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Manage traefik service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.traefik : Restart traefik service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Deploy manager service] **************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply manager role] ******************************************************",
                                "included: osism.services.manager for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/ansible)",
                                "changed: [testbed-manager] => (item=/opt/archive)",
                                "changed: [testbed-manager] => (item=/opt/manager/configuration)",
                                "changed: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "changed: [testbed-manager] => (item=/opt/manager/secrets)",
                                "changed: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "changed: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "changed: [testbed-manager] => (item=ara)",
                                "changed: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "changed: [testbed-manager] => (item=conductor)",
                                "changed: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "changed: [testbed-manager] => (item=osism)",
                                "changed: [testbed-manager] => (item=osism-update-docker)",
                                "changed: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "changed: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:2025.1\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:2025.1\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   ",
                                "",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:2025.1             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "manager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    2 minutes ago        Up About a minute (healthy)   192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             2 minutes ago        Up About a minute (healthy)   8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   2 minutes ago        Up About a minute (healthy)   ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 2 minutes ago        Up About a minute (healthy)   ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 57 seconds (healthy)       ",
                                "manager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               2 minutes ago        Up About a minute (healthy)   ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                2 minutes ago        Up About a minute (healthy)   3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              2 minutes ago        Up About a minute (healthy)   ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  2 minutes ago        Up About a minute (healthy)   6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               2 minutes ago        Up About a minute             192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   ",
                                "osismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            2 minutes ago        Up About a minute (healthy)   ",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg",
                                "+ osism apply resolvconf -l testbed-manager",
                                "\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m",
                                "\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87d64233-f8c7-4af4-b070-9f08313cd5a9 (resolvconf) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:23:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87d64233-f8c7-4af4-b070-9f08313cd5a9 (resolvconf) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role resolvconf] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Friday 13 March 2026  00:23:17 +0000 (0:00:00.110)       0:00:00.110 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Friday 13 March 2026  00:23:20 +0000 (0:00:03.270)       0:00:03.380 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Friday 13 March 2026  00:23:20 +0000 (0:00:00.053)       0:00:03.434 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Friday 13 March 2026  00:23:20 +0000 (0:00:00.077)       0:00:03.512 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Friday 13 March 2026  00:23:21 +0000 (0:00:00.069)       0:00:03.581 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Friday 13 March 2026  00:23:21 +0000 (0:00:00.853)       0:00:04.435 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Friday 13 March 2026  00:23:21 +0000 (0:00:00.066)       0:00:04.501 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Friday 13 March 2026  00:23:22 +0000 (0:00:00.471)       0:00:04.972 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Friday 13 March 2026  00:23:22 +0000 (0:00:00.079)       0:00:05.052 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Friday 13 March 2026  00:23:23 +0000 (0:00:00.542)       0:00:05.594 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Friday 13 March 2026  00:23:24 +0000 (0:00:01.026)       0:00:06.621 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Friday 13 March 2026  00:23:26 +0000 (0:00:01.962)       0:00:08.583 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Friday 13 March 2026  00:23:26 +0000 (0:00:00.080)       0:00:08.664 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:23:27 +0000 (0:00:01.116)       0:00:09.780 ********** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.27s",
                                "osism.commons.resolvconf : Start/enable systemd-resolved service -------- 1.96s",
                                "osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.12s",
                                "osism.commons.resolvconf : Copy configuration files --------------------- 1.03s",
                                "osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 0.85s",
                                "osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.54s",
                                "osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.47s",
                                "osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s",
                                "osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.08s",
                                "osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.08s",
                                "osism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s",
                                "osism.commons.resolvconf : Install package systemd-resolved ------------- 0.07s",
                                "osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.05s",
                                "+ osism apply sshconfig",
                                "\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m",
                                "\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f3d8db8b-09c0-4107-8621-a25ba374ce63 (sshconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:23:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f3d8db8b-09c0-4107-8621-a25ba374ce63 (sshconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role sshconfig] ****************************************************",
                                "",
                                "TASK [osism.commons.sshconfig : Get home directory of operator user] ***********",
                                "Friday 13 March 2026  00:23:43 +0000 (0:00:00.160)       0:00:00.160 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************",
                                "Friday 13 March 2026  00:23:44 +0000 (0:00:00.551)       0:00:00.711 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure config for each host exist] *************",
                                "Friday 13 March 2026  00:23:45 +0000 (0:00:00.498)       0:00:01.210 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Add extra config] ******************************",
                                "Friday 13 March 2026  00:23:50 +0000 (0:00:05.494)       0:00:06.705 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Assemble ssh config] ***************************",
                                "Friday 13 March 2026  00:23:50 +0000 (0:00:00.057)       0:00:06.762 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:23:51 +0000 (0:00:00.525)       0:00:07.288 ********** ",
                                "=============================================================================== ",
                                "osism.commons.sshconfig : Ensure config for each host exist ------------- 5.49s",
                                "osism.commons.sshconfig : Get home directory of operator user ----------- 0.55s",
                                "osism.commons.sshconfig : Assemble ssh config --------------------------- 0.53s",
                                "osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.50s",
                                "osism.commons.sshconfig : Add extra config ------------------------------ 0.06s",
                                "+ osism apply known-hosts",
                                "\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m",
                                "\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9eb0a440-9b1a-4fed-95c7-27da1b6b3a51 (known-hosts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:24:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9eb0a440-9b1a-4fed-95c7-27da1b6b3a51 (known-hosts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role known_hosts] **************************************************",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***",
                                "Friday 13 March 2026  00:24:07 +0000 (0:00:00.161)       0:00:00.161 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***",
                                "Friday 13 March 2026  00:24:13 +0000 (0:00:05.868)       0:00:06.029 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:13 +0000 (0:00:00.164)       0:00:06.194 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJuy8jCWnaLTBq9iEMruLjIH5rb6Sdsk/mdqEoBrT+Yh)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDUjn3pbkhTVDSfIvFWyrTPSAlrQ2qEHx+cKiucmRnylN8/dOPmVnzIWXKglSPdWPk0RXKjrAqkgk3asrjIyONAMu+Yodj9dMm7cQyJugykZzlsxrrqf29B1jllI0Vv0CmG+0aNOeySPU9lAjyCBT5bsvRHvFy6snz+unXWZ2P7ptsbZe5u7K5Gpl+5zsQpESQ/mvU2X1Rtvgn+7kxrwa0meG+JS/WYnluFbB1UBpYaaynx+6afbQBzJmSHJPU1PbbcVumyIvkBj2XL4ivi7Hi4HDtjSyICV22yl2/gSe6TrqPpRU/pAIy+w2eBb02FPcrt0fPb4UXUN1wqFPBSXYpA2UA89dw/lHJSTZNfCggEAEWXL3QGap2HTSiDliiowmuogw3hXL3l/M4RmA+CWB9V/sLY9ziYWc5p+3PBYs7SZCCQUy0nRFs82CqxQdBUnYu5ZJKZ8Jvo2ciXijV47faOxkEKDV5ur6sIYUuHJ3Cs5tPhSLFNL4l5Go2owYA8Sok=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCa5ZI05+fptRQNYqGW5ymCBzvxotWxdg4u9MPllpnlgiAaEzku1jQXOnCFnRCqR4no9Bp6IP/XbXsA6cIlx+cg=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:14 +0000 (0:00:01.164)       0:00:07.358 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBlWLCAhw75OFUH+WfJRsWoE7r4lRId0a5SuaQcF3RCn77R6tGXubcwvekH3i36AYoPVZ1f28/in+9qmOfBZEdg=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCx2OXD5KvuxQQvuuk/ZR3TYhjd2oCUgvGGlLl59UdbEWb6avU9P2aOZn394QvLxl/W0O5CVw+zcEx4rci/P5nYCHHzHb2uro0bRaBfdqQ0B4uqjFCIO8/aIULyDiBg0pDyxW6GJI0qERaXhOzc8z9bbyh9ve7zxqRm4M8cexuc+/79N/TqmUTbg+vQTTGJrHcKQrHFxIuBoH6NLeQaLRXfl6mb+2NmKp1bHHnrB0/QCBs6ry0PyJAgH6cLlD4sAiU8vC4IAkOVJu8tH6xA5ADISAS/SYx7M0F2+Ai9zUg3m+z3Kx0Ye9ZX+fyIeBm+dF9BV4/w6sSGqA/7L92pthy1dvpMADYsbNGL2graufYaxXiDCy4sGIzzPnzQCimpmIPkB7btUi/Fg6lzIG+RgHEyabKa++kbUPALty/qKm/ZLWsEwzoIvpMIZ8thOCHmYjHAYixR0kJO50v+A4udUDjmolwsMz8F9RCtxvxdgbHCN0MPKHNTNFENF/MPjLtOgXE=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINFXmFixrYsGuD6iK/5RZVMg1Bk/21IF1SHB4JPBI/fc)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:15 +0000 (0:00:00.975)       0:00:08.333 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ893j22t3cuMmiIB4zlWJMjyKFbTwJcc0BPH2rFyRTcNsdUzljqJpt4If0YzK+kUx9KSF9+7IQYQ5r+GlO4w+HloWfyBBsrzV/lf1w3dB5vNG9fhbRwOfegY+OI51PhSsJz2aZcmgI9X7Dy7M/9OQs9qQFxwjdx4OGwhDu0j+MwYiAUMLmju52inCVoEte9gr6XrFDex3kQA9oZVsW886Milfpj7a/kI31O85i3+RkrS1zWa064cZmnN/vIiJnhjezoSIGD/RHqIJS/k3GCnJTNMX+4WNRbjWPoALbiSXZl/WnofheFKuV9MSa+ylX1xIk1ua1E4cgCzI1m68m3909ufoY2T0YIQahM9iifMT664yLltMZqtujwRdyqbKL1+Xh6Fu+0REtnZgi50dHRR0xDzYfGTkjGab7V61ma2lhvMtpDg389lQwXFnUeKgpYZ2TfLIkdmu6c0q5rwPkrPk2jl3hJ6yjXLlKLtZ/xVL6FLrgd+n3xXkWpEUNm66WWk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIO25puKqGyzZ0nxVExnHcaYjmw7grNnYJGgIjtxandl1sKw4yHddITZ96S2Xx34RmTga6QLQnkmN6cp3DjcIYw=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEx3YF+TLygLxtcQFHnbHVeSvusJcGQANr1LDnZFBER1)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:16 +0000 (0:00:01.014)       0:00:09.347 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJyptwywAKmXERvxIbipTGF0jGA3OcJIXuRaNN4oaasRFcHs6TIsI+eCRt0U0JeTo2gQls1444wdlWE0zcBq5hw=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDj8+jynPJOfpKxwEdGVPOTygQAdAZ45l3zF2eb0Z7tXue7wdHzMCuQbVDXN1e6hHJYWw3WLeOiNkdZ1jQRvGehGRX9mI0GV+g4Pxx3t/fxT7IOxiw5Vmc1w6uB/vWWEum2fTT//tmBToA+ZQK8BzbjY9xum4PsEz+frApichK3SkheQJYJFKSinttaw39//rEMu9f6ZZaKBKmQFRnWANQnEaBUsAGP8BrOHJnc9JorlecS0qnVKDXLjX36qOTUj7CG89fnZGx4hmra6BWUtqGMLbiZmmKADImxzyC1NMvlYoECnt2vepEC+vK8dV0ITuCTaGGvlYrDRFOFaBn+mWGcwqLsMy1uLzKk3eKB8hZJ/6lh8AqO6oygvJU/tBDpbadFVBxHRq8qXyoe0Lrtnncd4jUEcx4SrBzHO+HjKnStSHxY93VtYRDLsP6rc4utNl3wKk1VbXHzZ7NdssFE/+uznzPcI18xp7knIn8BCIGkIUd8gNJtw+XUiEdgdd9OARM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPXUTKHySIJuePlD23IvyWC7WWIcXkfHppc5thZrV4EL)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:17 +0000 (0:00:01.014)       0:00:10.362 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVi8GbWLEVnN2cq1UtQFymBGqLAadzfhhMKJN4hSR6qu1xpVVoC8m4wmpU/M0ryTlbp8HiNhe/5fHTQiRkite0mThpP9gcYDkXsJsJyIWhSZ+w9d1KdZy33iHVLFmgHWhTmCa3Zz9ptJDVZWyTSqIf9FV5RQJDVO4F43bf4Oz/ZKIKZGV+pb81Nl21s+OCcr3PnCchaKMgRHuUiz1wlaaE5gJeyS3muh1jpc9jxi47IAqlvbdWsP0n2iFFy2Vz+FSd+VcRoHL/4EnbAUMs8278wJlL8bcpGWH3qls69gc8vQRsUmH7HgZn8yC9CHTazbShsy8y3eXWOXyiDPQnmDlJVDr1fAWnBPVCqfo3gWvURBXbG8/MdLKfZmCDpYouL9T8UR6Y2SdQXtGFpx0L4Hb7Oloi8EiUt1d8k86Tqol7MxKL1v3VBct+8WVQUrwhtApgYOLT2XXSwuMb6wiKzuLswvyUOZzMoXipWboEoUl74mwoe17QSROul5/otMlzB08=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJcatKxhtat8HayXurSHkcdRyxEd3QAmE1LHT07qM0cl)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAW1cyPz5B71lNK9YON6FQwdjfZdu5EuN9ADHhJNE8uj72sCkjktOPZrPm9burSJUqBE9/+Ran1oU0b/+jtRqpA=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:18 +0000 (0:00:01.007)       0:00:11.370 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG+ZNRyzVJ0g92UyubSMydcqVN3BYyGRL3kj7QhcmgiEjpPyLxZ9nUsQ5jNAYVIdFk2aW7qkU7iitX5aRU9djz4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCrdcb4tS416Fq7fSiHkjU70NV58/hXHf3StBiWpNG/pF4gjnQjbaj31jEUY9Os6rniNll/iQ7pI4HG9pnZ3UOWjg7hk9VBz+lod9Fb8RrA4C4/WdJo20YR7u5RqlaQBSMnZksbvzaZCOdxqDF6BDNg3Eq1x149dqcyfCaoQinLS2zAHPZofL6VuvRib33wscojz7mZZQkgMfpty2ensqfW6jypAtaD1fz87wUb2Eu7c1oedv8DQlaYPCoAPWmdIuLpmcxAyK0jFYVf51mrWFeE3KV4wKDLwRU8fm8B4ciUFP8QkOXH70vx3NT2nc2zORhJAI9gqYWBg3LS+LGZL0Zm3cCy0JmyKeP6iN0X9Vz11T0EL1IkQj62xN6y6yG1ArZ46gNDfY5997YL7PbxGtD5840z0Ap2A8JPtu7eAOUuRfvAoDrOLdDqb6glRYeCMMKWEgpSFRdSIfoe/S4dx/3eDbZmm2ajcm/B8cO0fKiu9n96jP0hSQNcwtQUcl0drG8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN5nXX6lrkdkqdrcnFbDBZcznK46Ot6IznDrG1/JbJtR)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:19 +0000 (0:00:00.994)       0:00:12.365 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEq41I4ETswmUujQlL1aob78tYUYDVmH1bCbKnsi1rza)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBVzQyOT4WunW9bbwQfaHK8PAR6yp3J9G8oqLno+QCnGild5O03duxDMS2WWW+12cOjjlipXQEOWeckreVtIjBOlvfLkRpAbalcOHRAYO2ZD6+IhLX4zsCc+b5N+j6uGSjOlG2P/i9LIRJgyKeZoPdQnxRhQWmY9CP+duDs5jutmOZuzSDimO/S0gHdCqPJ+c9FsxX7AA3gZFBTjEqvHB8SA+hMKu3GCnDbhmx1FourmBF0E+2EQONrNazFmHZHEP8IxrLk4jRIwQ37N2Opoay913z0w4YvwCkcU1oh+fj0j8ggSwfKB0s62HA/qN9/p6z315oHJQYoJyr8+5+oyIS0piVUYORJrOIcpMdnIWepAJILND6E8OhX5YqFf06+nmf8y33tBFFJMziN52jm3fJZAeFR5vtVw4Uy24tixMHYY8IETvBrWxBwO13wDmeAAORfV7KHYQgdcwWbZlHDe1xnFNJ9v95FVaTuH2q7ceq4oAMFfnXRWRJgVxQH37WOXc=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMClF11FAZYr5fRF50WO1AWC8MiLmVAfUBX2NUtUP9tf987+Msed8myo78FC7NXXIO10EyVDIODNmrOuQXiJJ3A=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***",
                                "Friday 13 March 2026  00:24:20 +0000 (0:00:00.978)       0:00:13.343 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***",
                                "Friday 13 March 2026  00:24:25 +0000 (0:00:05.230)       0:00:18.574 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:26 +0000 (0:00:00.171)       0:00:18.745 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJuy8jCWnaLTBq9iEMruLjIH5rb6Sdsk/mdqEoBrT+Yh)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDUjn3pbkhTVDSfIvFWyrTPSAlrQ2qEHx+cKiucmRnylN8/dOPmVnzIWXKglSPdWPk0RXKjrAqkgk3asrjIyONAMu+Yodj9dMm7cQyJugykZzlsxrrqf29B1jllI0Vv0CmG+0aNOeySPU9lAjyCBT5bsvRHvFy6snz+unXWZ2P7ptsbZe5u7K5Gpl+5zsQpESQ/mvU2X1Rtvgn+7kxrwa0meG+JS/WYnluFbB1UBpYaaynx+6afbQBzJmSHJPU1PbbcVumyIvkBj2XL4ivi7Hi4HDtjSyICV22yl2/gSe6TrqPpRU/pAIy+w2eBb02FPcrt0fPb4UXUN1wqFPBSXYpA2UA89dw/lHJSTZNfCggEAEWXL3QGap2HTSiDliiowmuogw3hXL3l/M4RmA+CWB9V/sLY9ziYWc5p+3PBYs7SZCCQUy0nRFs82CqxQdBUnYu5ZJKZ8Jvo2ciXijV47faOxkEKDV5ur6sIYUuHJ3Cs5tPhSLFNL4l5Go2owYA8Sok=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBCa5ZI05+fptRQNYqGW5ymCBzvxotWxdg4u9MPllpnlgiAaEzku1jQXOnCFnRCqR4no9Bp6IP/XbXsA6cIlx+cg=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:27 +0000 (0:00:01.024)       0:00:19.769 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINFXmFixrYsGuD6iK/5RZVMg1Bk/21IF1SHB4JPBI/fc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCx2OXD5KvuxQQvuuk/ZR3TYhjd2oCUgvGGlLl59UdbEWb6avU9P2aOZn394QvLxl/W0O5CVw+zcEx4rci/P5nYCHHzHb2uro0bRaBfdqQ0B4uqjFCIO8/aIULyDiBg0pDyxW6GJI0qERaXhOzc8z9bbyh9ve7zxqRm4M8cexuc+/79N/TqmUTbg+vQTTGJrHcKQrHFxIuBoH6NLeQaLRXfl6mb+2NmKp1bHHnrB0/QCBs6ry0PyJAgH6cLlD4sAiU8vC4IAkOVJu8tH6xA5ADISAS/SYx7M0F2+Ai9zUg3m+z3Kx0Ye9ZX+fyIeBm+dF9BV4/w6sSGqA/7L92pthy1dvpMADYsbNGL2graufYaxXiDCy4sGIzzPnzQCimpmIPkB7btUi/Fg6lzIG+RgHEyabKa++kbUPALty/qKm/ZLWsEwzoIvpMIZ8thOCHmYjHAYixR0kJO50v+A4udUDjmolwsMz8F9RCtxvxdgbHCN0MPKHNTNFENF/MPjLtOgXE=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBlWLCAhw75OFUH+WfJRsWoE7r4lRId0a5SuaQcF3RCn77R6tGXubcwvekH3i36AYoPVZ1f28/in+9qmOfBZEdg=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:28 +0000 (0:00:01.030)       0:00:20.800 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEx3YF+TLygLxtcQFHnbHVeSvusJcGQANr1LDnZFBER1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ893j22t3cuMmiIB4zlWJMjyKFbTwJcc0BPH2rFyRTcNsdUzljqJpt4If0YzK+kUx9KSF9+7IQYQ5r+GlO4w+HloWfyBBsrzV/lf1w3dB5vNG9fhbRwOfegY+OI51PhSsJz2aZcmgI9X7Dy7M/9OQs9qQFxwjdx4OGwhDu0j+MwYiAUMLmju52inCVoEte9gr6XrFDex3kQA9oZVsW886Milfpj7a/kI31O85i3+RkrS1zWa064cZmnN/vIiJnhjezoSIGD/RHqIJS/k3GCnJTNMX+4WNRbjWPoALbiSXZl/WnofheFKuV9MSa+ylX1xIk1ua1E4cgCzI1m68m3909ufoY2T0YIQahM9iifMT664yLltMZqtujwRdyqbKL1+Xh6Fu+0REtnZgi50dHRR0xDzYfGTkjGab7V61ma2lhvMtpDg389lQwXFnUeKgpYZ2TfLIkdmu6c0q5rwPkrPk2jl3hJ6yjXLlKLtZ/xVL6FLrgd+n3xXkWpEUNm66WWk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIO25puKqGyzZ0nxVExnHcaYjmw7grNnYJGgIjtxandl1sKw4yHddITZ96S2Xx34RmTga6QLQnkmN6cp3DjcIYw=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:29 +0000 (0:00:00.983)       0:00:21.784 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDj8+jynPJOfpKxwEdGVPOTygQAdAZ45l3zF2eb0Z7tXue7wdHzMCuQbVDXN1e6hHJYWw3WLeOiNkdZ1jQRvGehGRX9mI0GV+g4Pxx3t/fxT7IOxiw5Vmc1w6uB/vWWEum2fTT//tmBToA+ZQK8BzbjY9xum4PsEz+frApichK3SkheQJYJFKSinttaw39//rEMu9f6ZZaKBKmQFRnWANQnEaBUsAGP8BrOHJnc9JorlecS0qnVKDXLjX36qOTUj7CG89fnZGx4hmra6BWUtqGMLbiZmmKADImxzyC1NMvlYoECnt2vepEC+vK8dV0ITuCTaGGvlYrDRFOFaBn+mWGcwqLsMy1uLzKk3eKB8hZJ/6lh8AqO6oygvJU/tBDpbadFVBxHRq8qXyoe0Lrtnncd4jUEcx4SrBzHO+HjKnStSHxY93VtYRDLsP6rc4utNl3wKk1VbXHzZ7NdssFE/+uznzPcI18xp7knIn8BCIGkIUd8gNJtw+XUiEdgdd9OARM=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJyptwywAKmXERvxIbipTGF0jGA3OcJIXuRaNN4oaasRFcHs6TIsI+eCRt0U0JeTo2gQls1444wdlWE0zcBq5hw=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPXUTKHySIJuePlD23IvyWC7WWIcXkfHppc5thZrV4EL)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:30 +0000 (0:00:01.006)       0:00:22.790 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAW1cyPz5B71lNK9YON6FQwdjfZdu5EuN9ADHhJNE8uj72sCkjktOPZrPm9burSJUqBE9/+Ran1oU0b/+jtRqpA=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDVi8GbWLEVnN2cq1UtQFymBGqLAadzfhhMKJN4hSR6qu1xpVVoC8m4wmpU/M0ryTlbp8HiNhe/5fHTQiRkite0mThpP9gcYDkXsJsJyIWhSZ+w9d1KdZy33iHVLFmgHWhTmCa3Zz9ptJDVZWyTSqIf9FV5RQJDVO4F43bf4Oz/ZKIKZGV+pb81Nl21s+OCcr3PnCchaKMgRHuUiz1wlaaE5gJeyS3muh1jpc9jxi47IAqlvbdWsP0n2iFFy2Vz+FSd+VcRoHL/4EnbAUMs8278wJlL8bcpGWH3qls69gc8vQRsUmH7HgZn8yC9CHTazbShsy8y3eXWOXyiDPQnmDlJVDr1fAWnBPVCqfo3gWvURBXbG8/MdLKfZmCDpYouL9T8UR6Y2SdQXtGFpx0L4Hb7Oloi8EiUt1d8k86Tqol7MxKL1v3VBct+8WVQUrwhtApgYOLT2XXSwuMb6wiKzuLswvyUOZzMoXipWboEoUl74mwoe17QSROul5/otMlzB08=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJcatKxhtat8HayXurSHkcdRyxEd3QAmE1LHT07qM0cl)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:31 +0000 (0:00:00.974)       0:00:23.765 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG+ZNRyzVJ0g92UyubSMydcqVN3BYyGRL3kj7QhcmgiEjpPyLxZ9nUsQ5jNAYVIdFk2aW7qkU7iitX5aRU9djz4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCrdcb4tS416Fq7fSiHkjU70NV58/hXHf3StBiWpNG/pF4gjnQjbaj31jEUY9Os6rniNll/iQ7pI4HG9pnZ3UOWjg7hk9VBz+lod9Fb8RrA4C4/WdJo20YR7u5RqlaQBSMnZksbvzaZCOdxqDF6BDNg3Eq1x149dqcyfCaoQinLS2zAHPZofL6VuvRib33wscojz7mZZQkgMfpty2ensqfW6jypAtaD1fz87wUb2Eu7c1oedv8DQlaYPCoAPWmdIuLpmcxAyK0jFYVf51mrWFeE3KV4wKDLwRU8fm8B4ciUFP8QkOXH70vx3NT2nc2zORhJAI9gqYWBg3LS+LGZL0Zm3cCy0JmyKeP6iN0X9Vz11T0EL1IkQj62xN6y6yG1ArZ46gNDfY5997YL7PbxGtD5840z0Ap2A8JPtu7eAOUuRfvAoDrOLdDqb6glRYeCMMKWEgpSFRdSIfoe/S4dx/3eDbZmm2ajcm/B8cO0fKiu9n96jP0hSQNcwtQUcl0drG8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN5nXX6lrkdkqdrcnFbDBZcznK46Ot6IznDrG1/JbJtR)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Friday 13 March 2026  00:24:32 +0000 (0:00:01.008)       0:00:24.773 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEq41I4ETswmUujQlL1aob78tYUYDVmH1bCbKnsi1rza)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDBVzQyOT4WunW9bbwQfaHK8PAR6yp3J9G8oqLno+QCnGild5O03duxDMS2WWW+12cOjjlipXQEOWeckreVtIjBOlvfLkRpAbalcOHRAYO2ZD6+IhLX4zsCc+b5N+j6uGSjOlG2P/i9LIRJgyKeZoPdQnxRhQWmY9CP+duDs5jutmOZuzSDimO/S0gHdCqPJ+c9FsxX7AA3gZFBTjEqvHB8SA+hMKu3GCnDbhmx1FourmBF0E+2EQONrNazFmHZHEP8IxrLk4jRIwQ37N2Opoay913z0w4YvwCkcU1oh+fj0j8ggSwfKB0s62HA/qN9/p6z315oHJQYoJyr8+5+oyIS0piVUYORJrOIcpMdnIWepAJILND6E8OhX5YqFf06+nmf8y33tBFFJMziN52jm3fJZAeFR5vtVw4Uy24tixMHYY8IETvBrWxBwO13wDmeAAORfV7KHYQgdcwWbZlHDe1xnFNJ9v95FVaTuH2q7ceq4oAMFfnXRWRJgVxQH37WOXc=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMClF11FAZYr5fRF50WO1AWC8MiLmVAfUBX2NUtUP9tf987+Msed8myo78FC7NXXIO10EyVDIODNmrOuQXiJJ3A=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write static known_hosts entries] ************",
                                "Friday 13 March 2026  00:24:33 +0000 (0:00:01.050)       0:00:25.824 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write extra known_hosts entries] *************",
                                "Friday 13 March 2026  00:24:33 +0000 (0:00:00.155)       0:00:25.979 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Delete known_hosts entries] ******************",
                                "Friday 13 March 2026  00:24:33 +0000 (0:00:00.046)       0:00:26.026 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Set file permissions] ************************",
                                "Friday 13 March 2026  00:24:33 +0000 (0:00:00.041)       0:00:26.067 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:24:34 +0000 (0:00:00.701)       0:00:26.768 ********** ",
                                "=============================================================================== ",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.87s",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.23s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.16s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.99s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.97s",
                                "osism.commons.known_hosts : Set file permissions ------------------------ 0.70s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s",
                                "osism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s",
                                "+ osism apply squid",
                                "\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m",
                                "\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 076722ab-df37-4595-91ef-2a9c0563a707 (squid) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:24:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 076722ab-df37-4595-91ef-2a9c0563a707 (squid) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role squid] ********************************************************",
                                "",
                                "TASK [osism.services.squid : Include install tasks] ****************************",
                                "Friday 13 March 2026  00:24:50 +0000 (0:00:00.119)       0:00:00.119 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Install required packages] ************************",
                                "Friday 13 March 2026  00:24:50 +0000 (0:00:00.063)       0:00:00.183 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Create required directories] **********************",
                                "Friday 13 March 2026  00:24:51 +0000 (0:00:01.011)       0:00:01.194 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy squid configuration files] *******************",
                                "Friday 13 March 2026  00:24:52 +0000 (0:00:01.004)       0:00:02.199 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***",
                                "Friday 13 March 2026  00:24:53 +0000 (0:00:00.953)       0:00:03.153 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy docker-compose.yml file] *********************",
                                "Friday 13 March 2026  00:24:53 +0000 (0:00:00.318)       0:00:03.471 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Manage squid service] *****************************",
                                "Friday 13 March 2026  00:24:54 +0000 (0:00:00.789)       0:00:04.261 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Restart squid service] *****************",
                                "Friday 13 March 2026  00:25:24 +0000 (0:00:29.893)       0:00:34.155 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******",
                                "Friday 13 March 2026  00:25:36 +0000 (0:00:11.926)       0:00:46.081 ********** ",
                                "Pausing for 60 seconds",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***",
                                "Friday 13 March 2026  00:26:36 +0000 (0:01:00.084)       0:01:46.165 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****",
                                "Friday 13 March 2026  00:26:36 +0000 (0:00:00.059)       0:01:46.224 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:26:36 +0000 (0:00:00.597)       0:01:46.821 ********** ",
                                "=============================================================================== ",
                                "osism.services.squid : Wait for squid service to start ----------------- 60.08s",
                                "osism.services.squid : Manage squid service ---------------------------- 29.89s",
                                "osism.services.squid : Restart squid service --------------------------- 11.93s",
                                "osism.services.squid : Install required packages ------------------------ 1.01s",
                                "osism.services.squid : Create required directories ---------------------- 1.00s",
                                "osism.services.squid : Copy squid configuration files ------------------- 0.95s",
                                "osism.services.squid : Copy docker-compose.yml file --------------------- 0.79s",
                                "osism.services.squid : Wait for an healthy squid service ---------------- 0.60s",
                                "osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.32s",
                                "osism.services.squid : Include install tasks ---------------------------- 0.06s",
                                "osism.services.squid : Register that squid service was restarted -------- 0.06s",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla",
                                "+ set -e",
                                "+ NAMESPACE=kolla",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "++ semver latest 9.0.0",
                                "+ [[ -1 -lt 0 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply operator -u ubuntu -l testbed-nodes",
                                "\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m",
                                "\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a138e641-c2e4-42c8-867d-347889857868 (operator) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:26:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task a138e641-c2e4-42c8-867d-347889857868 (operator) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Make ssh pipelining working] *********************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Friday 13 March 2026  00:26:53 +0000 (0:00:00.101)       0:00:00.101 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Do not require tty for all users] ****************************************",
                                "Friday 13 March 2026  00:26:57 +0000 (0:00:04.234)       0:00:04.336 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role operator] *****************************************************",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "Friday 13 March 2026  00:26:58 +0000 (0:00:00.948)       0:00:05.285 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "Friday 13 March 2026  00:26:58 +0000 (0:00:00.145)       0:00:05.431 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "Friday 13 March 2026  00:26:58 +0000 (0:00:00.165)       0:00:05.596 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "Friday 13 March 2026  00:26:59 +0000 (0:00:00.645)       0:00:06.242 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "Friday 13 March 2026  00:27:00 +0000 (0:00:00.953)       0:00:07.195 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "Friday 13 March 2026  00:27:01 +0000 (0:00:01.242)       0:00:08.437 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "Friday 13 March 2026  00:27:02 +0000 (0:00:01.323)       0:00:09.760 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m",
                                "\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m",
                                "\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "Friday 13 March 2026  00:27:04 +0000 (0:00:01.280)       0:00:11.042 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "Friday 13 March 2026  00:27:04 +0000 (0:00:00.134)       0:00:11.176 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "Friday 13 March 2026  00:27:04 +0000 (0:00:00.165)       0:00:11.342 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "Friday 13 March 2026  00:27:05 +0000 (0:00:00.656)       0:00:11.998 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "Friday 13 March 2026  00:27:05 +0000 (0:00:00.164)       0:00:12.163 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "Friday 13 March 2026  00:27:05 +0000 (0:00:00.731)       0:00:12.894 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "Friday 13 March 2026  00:27:06 +0000 (0:00:00.129)       0:00:13.024 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "Friday 13 March 2026  00:27:06 +0000 (0:00:00.145)       0:00:13.169 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "Friday 13 March 2026  00:27:06 +0000 (0:00:00.141)       0:00:13.311 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "Friday 13 March 2026  00:27:07 +0000 (0:00:00.701)       0:00:14.012 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:27:07 +0000 (0:00:00.214)       0:00:14.226 ********** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 4.23s",
                                "osism.commons.operator : Copy user sudoers file ------------------------- 1.32s",
                                "osism.commons.operator : Set language variables in .bashrc configuration file --- 1.28s",
                                "osism.commons.operator : Add user to additional groups ------------------ 1.24s",
                                "osism.commons.operator : Create user ------------------------------------ 0.95s",
                                "Do not require tty for all users ---------------------------------------- 0.95s",
                                "osism.commons.operator : Set ssh authorized keys ------------------------ 0.73s",
                                "osism.commons.operator : Set password ----------------------------------- 0.70s",
                                "osism.commons.operator : Create .ssh directory -------------------------- 0.66s",
                                "osism.commons.operator : Create operator group -------------------------- 0.65s",
                                "osism.commons.operator : Unset & lock password -------------------------- 0.21s",
                                "osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.17s",
                                "osism.commons.operator : Set operator_groups variable to default value --- 0.17s",
                                "osism.commons.operator : Check number of SSH authorized keys ------------ 0.16s",
                                "osism.commons.operator : Gather variables for each operating system ----- 0.15s",
                                "osism.commons.operator : Set authorized GitHub accounts ----------------- 0.15s",
                                "osism.commons.operator : Delete authorized GitHub accounts -------------- 0.14s",
                                "osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.13s",
                                "osism.commons.operator : Delete ssh authorized keys --------------------- 0.13s",
                                "+ osism apply --environment custom facts",
                                "\u001b[32m2026-03-13 00:27:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m",
                                "\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7bcd12de-b383-4c91-a9da-2851d605c017 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:27:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7bcd12de-b383-4c91-a9da-2851d605c017 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy custom network devices fact] ****************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Friday 13 March 2026  00:27:23 +0000 (0:00:00.052)       0:00:00.052 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Copy fact file] **********************************************************",
                                "Friday 13 March 2026  00:27:24 +0000 (0:00:01.255)       0:00:01.307 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Copy custom ceph devices facts] ******************************************",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Friday 13 March 2026  00:27:25 +0000 (0:00:01.018)       0:00:02.326 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Friday 13 March 2026  00:27:25 +0000 (0:00:00.087)       0:00:02.414 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Friday 13 March 2026  00:27:25 +0000 (0:00:00.164)       0:00:02.578 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Friday 13 March 2026  00:27:25 +0000 (0:00:00.177)       0:00:02.756 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Friday 13 March 2026  00:27:26 +0000 (0:00:00.097)       0:00:02.853 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Friday 13 March 2026  00:27:26 +0000 (0:00:00.439)       0:00:03.293 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Friday 13 March 2026  00:27:26 +0000 (0:00:00.086)       0:00:03.379 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Friday 13 March 2026  00:27:27 +0000 (0:00:01.005)       0:00:04.384 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Friday 13 March 2026  00:27:28 +0000 (0:00:00.438)       0:00:04.823 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Friday 13 March 2026  00:27:29 +0000 (0:00:01.086)       0:00:05.909 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Install required packages (RedHat)] **************************************",
                                "Friday 13 March 2026  00:27:46 +0000 (0:00:17.755)       0:00:23.664 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (Debian)] **************************************",
                                "Friday 13 March 2026  00:27:46 +0000 (0:00:00.103)       0:00:23.768 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Friday 13 March 2026  00:27:56 +0000 (0:00:09.147)       0:00:32.915 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "Friday 13 March 2026  00:27:56 +0000 (0:00:00.442)       0:00:33.358 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Friday 13 March 2026  00:28:00 +0000 (0:00:03.739)       0:00:37.097 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:28:01 +0000 (0:00:01.587)       0:00:38.685 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:28:06 +0000 (0:00:04.840)       0:00:43.525 ********** ",
                                "=============================================================================== ",
                                "osism.commons.repository : Update package cache ------------------------ 17.76s",
                                "Install required packages (Debian) -------------------------------------- 9.15s",
                                "Gathers facts about hosts ----------------------------------------------- 4.84s",
                                "Copy fact files --------------------------------------------------------- 3.74s",
                                "osism.commons.repository : Force update of package cache ---------------- 1.59s",
                                "Create custom facts directory ------------------------------------------- 1.26s",
                                "osism.commons.repository : Copy ubuntu.sources file --------------------- 1.09s",
                                "Copy fact file ---------------------------------------------------------- 1.02s",
                                "osism.commons.repository : Copy 99osism apt configuration --------------- 1.01s",
                                "Create custom facts directory ------------------------------------------- 0.44s",
                                "osism.commons.repository : Remove sources.list file --------------------- 0.44s",
                                "osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.44s",
                                "osism.commons.repository : Set repositories to default ------------------ 0.18s",
                                "osism.commons.repository : Set repository_default fact to default value --- 0.16s",
                                "Install required packages (RedHat) -------------------------------------- 0.10s",
                                "osism.commons.repository : Include distribution specific repository tasks --- 0.10s",
                                "osism.commons.repository : Gather variables for each operating system --- 0.09s",
                                "osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.09s",
                                "+ osism apply bootstrap",
                                "\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m",
                                "\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 62f251aa-bb4c-40fe-8fe8-9c5aca7a7d34 (bootstrap) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:28:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 62f251aa-bb4c-40fe-8fe8-9c5aca7a7d34 (bootstrap) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on state bootstrap] ************************************",
                                "",
                                "TASK [Group hosts based on state bootstrap] ************************************",
                                "Friday 13 March 2026  00:28:23 +0000 (0:00:00.137)       0:00:00.137 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:28:23 +0000 (0:00:00.229)       0:00:00.367 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts (if using --limit)] ***************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:28:27 +0000 (0:00:03.546)       0:00:03.914 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 1] ********************************************",
                                "",
                                "TASK [osism.commons.hostname : Set hostname] ***********************************",
                                "Friday 13 March 2026  00:28:27 +0000 (0:00:00.491)       0:00:04.405 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.hostname : Copy /etc/hostname] *****************************",
                                "Friday 13 March 2026  00:28:29 +0000 (0:00:01.257)       0:00:05.663 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Include type specific tasks] ***********************",
                                "Friday 13 March 2026  00:28:30 +0000 (0:00:01.401)       0:00:07.065 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Copy /etc/hosts file] ******************************",
                                "Friday 13 March 2026  00:28:30 +0000 (0:00:00.285)       0:00:07.351 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Include distribution specific tasks] ***************",
                                "Friday 13 March 2026  00:28:33 +0000 (0:00:02.859)       0:00:10.211 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Configure proxy parameters for apt] ****************",
                                "Friday 13 March 2026  00:28:33 +0000 (0:00:00.236)       0:00:10.447 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Set system wide settings in environment file] ******",
                                "Friday 13 March 2026  00:28:34 +0000 (0:00:01.106)       0:00:11.554 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Remove system wide settings in environment file] ***",
                                "Friday 13 March 2026  00:28:35 +0000 (0:00:00.587)       0:00:12.141 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Friday 13 March 2026  00:28:36 +0000 (0:00:00.546)       0:00:12.687 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Friday 13 March 2026  00:28:36 +0000 (0:00:00.215)       0:00:12.903 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Friday 13 March 2026  00:28:36 +0000 (0:00:00.253)       0:00:13.157 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Friday 13 March 2026  00:28:36 +0000 (0:00:00.356)       0:00:13.513 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Friday 13 March 2026  00:28:38 +0000 (0:00:01.467)       0:00:14.981 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Friday 13 March 2026  00:28:38 +0000 (0:00:00.227)       0:00:15.209 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Friday 13 March 2026  00:28:39 +0000 (0:00:00.516)       0:00:15.726 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Friday 13 March 2026  00:28:39 +0000 (0:00:00.235)       0:00:15.961 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Friday 13 March 2026  00:28:39 +0000 (0:00:00.592)       0:00:16.553 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Friday 13 March 2026  00:28:41 +0000 (0:00:01.147)       0:00:17.701 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Friday 13 March 2026  00:28:42 +0000 (0:00:01.047)       0:00:18.748 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Friday 13 March 2026  00:28:42 +0000 (0:00:00.301)       0:00:19.050 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Friday 13 March 2026  00:28:43 +0000 (0:00:01.417)       0:00:20.467 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Friday 13 March 2026  00:28:44 +0000 (0:00:00.208)       0:00:20.676 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Friday 13 March 2026  00:28:44 +0000 (0:00:00.217)       0:00:20.893 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Friday 13 March 2026  00:28:44 +0000 (0:00:00.219)       0:00:21.113 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Friday 13 March 2026  00:28:44 +0000 (0:00:00.266)       0:00:21.379 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Friday 13 March 2026  00:28:45 +0000 (0:00:00.553)       0:00:21.933 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Friday 13 March 2026  00:28:45 +0000 (0:00:00.226)       0:00:22.160 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Friday 13 March 2026  00:28:46 +0000 (0:00:01.022)       0:00:23.182 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Friday 13 March 2026  00:28:47 +0000 (0:00:00.618)       0:00:23.800 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Friday 13 March 2026  00:28:48 +0000 (0:00:01.138)       0:00:24.939 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Gather variables for each operating system] *****",
                                "Friday 13 March 2026  00:29:06 +0000 (0:00:18.642)       0:00:43.581 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****",
                                "Friday 13 March 2026  00:29:07 +0000 (0:00:00.249)       0:00:43.831 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***",
                                "Friday 13 March 2026  00:29:07 +0000 (0:00:00.203)       0:00:44.035 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include distribution specific install tasks] ****",
                                "Friday 13 March 2026  00:29:07 +0000 (0:00:00.201)       0:00:44.237 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Install rsyslog package] ************************",
                                "Friday 13 March 2026  00:29:07 +0000 (0:00:00.281)       0:00:44.518 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********",
                                "Friday 13 March 2026  00:29:09 +0000 (0:00:01.825)       0:00:46.343 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Manage rsyslog service] *************************",
                                "Friday 13 March 2026  00:29:10 +0000 (0:00:00.994)       0:00:47.338 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include fluentd tasks] **************************",
                                "Friday 13 March 2026  00:29:11 +0000 (0:00:00.848)       0:00:48.186 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***",
                                "Friday 13 March 2026  00:29:11 +0000 (0:00:00.290)       0:00:48.477 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include additional log server tasks] ************",
                                "Friday 13 March 2026  00:29:12 +0000 (0:00:01.144)       0:00:49.621 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include logrotate tasks] ************************",
                                "Friday 13 March 2026  00:29:13 +0000 (0:00:00.212)       0:00:49.834 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Ensure logrotate package is installed] **********",
                                "Friday 13 March 2026  00:29:13 +0000 (0:00:00.261)       0:00:50.096 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************",
                                "Friday 13 March 2026  00:29:15 +0000 (0:00:01.782)       0:00:51.878 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Install util-linux-extra package] ****************",
                                "Friday 13 March 2026  00:29:16 +0000 (0:00:01.052)       0:00:52.930 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Sync hardware clock] *****************************",
                                "Friday 13 March 2026  00:29:27 +0000 (0:00:11.359)       0:01:04.289 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.configfs : Start sys-kernel-config mount] ******************",
                                "Friday 13 March 2026  00:29:28 +0000 (0:00:00.996)       0:01:05.286 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Gather variables for each operating system] *****",
                                "Friday 13 March 2026  00:29:29 +0000 (0:00:00.896)       0:01:06.183 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***",
                                "Friday 13 March 2026  00:29:29 +0000 (0:00:00.197)       0:01:06.381 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Include distribution specific package tasks] ****",
                                "Friday 13 March 2026  00:29:29 +0000 (0:00:00.206)       0:01:06.587 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install needrestart package] ********************",
                                "Friday 13 March 2026  00:29:30 +0000 (0:00:00.274)       0:01:06.862 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set needrestart mode] ***************************",
                                "Friday 13 March 2026  00:29:32 +0000 (0:00:01.926)       0:01:08.788 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***",
                                "Friday 13 March 2026  00:29:32 +0000 (0:00:00.550)       0:01:09.339 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Update package cache] ***************************",
                                "Friday 13 March 2026  00:29:32 +0000 (0:00:00.190)       0:01:09.529 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download upgrade packages] **********************",
                                "Friday 13 March 2026  00:29:34 +0000 (0:00:01.272)       0:01:10.802 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Upgrade packages] *******************************",
                                "Friday 13 March 2026  00:29:36 +0000 (0:00:01.943)       0:01:12.746 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download required packages] *********************",
                                "Friday 13 March 2026  00:29:39 +0000 (0:00:03.217)       0:01:15.963 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install required packages] **********************",
                                "Friday 13 March 2026  00:30:16 +0000 (0:00:37.340)       0:01:53.304 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove useless packages from the cache] *********",
                                "Friday 13 March 2026  00:31:38 +0000 (0:01:21.612)       0:03:14.917 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove dependencies that are no longer required] ***",
                                "Friday 13 March 2026  00:31:40 +0000 (0:00:01.758)       0:03:16.675 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Include sysctl tasks] *****************************",
                                "Friday 13 March 2026  00:31:49 +0000 (0:00:09.951)       0:03:26.627 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********",
                                "Friday 13 March 2026  00:31:50 +0000 (0:00:00.424)       0:03:27.051 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************",
                                "Friday 13 March 2026  00:31:51 +0000 (0:00:00.661)       0:03:27.712 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************",
                                "Friday 13 March 2026  00:31:56 +0000 (0:00:05.914)       0:03:33.627 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************",
                                "Friday 13 March 2026  00:31:58 +0000 (0:00:01.492)       0:03:35.119 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on network] *****************",
                                "Friday 13 March 2026  00:31:59 +0000 (0:00:01.447)       0:03:36.567 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************",
                                "Friday 13 March 2026  00:32:01 +0000 (0:00:01.533)       0:03:38.100 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "",
                                "TASK [osism.commons.limits : Include limits tasks] *****************************",
                                "Friday 13 March 2026  00:32:02 +0000 (0:00:00.551)       0:03:38.652 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Populate service facts] *************************",
                                "Friday 13 March 2026  00:32:02 +0000 (0:00:00.280)       0:03:38.932 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Check services] *********************************",
                                "Friday 13 March 2026  00:32:07 +0000 (0:00:05.359)       0:03:44.292 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Start/enable required services] *****************",
                                "Friday 13 March 2026  00:32:07 +0000 (0:00:00.302)       0:03:44.595 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Include distribution specific configure tasks] ******",
                                "Friday 13 March 2026  00:32:09 +0000 (0:00:01.119)       0:03:45.714 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove update-motd package] *************************",
                                "Friday 13 March 2026  00:32:09 +0000 (0:00:00.387)       0:03:46.102 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************",
                                "Friday 13 March 2026  00:32:10 +0000 (0:00:01.276)       0:03:47.378 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Disable the dynamic motd-news service] **************",
                                "Friday 13 March 2026  00:32:11 +0000 (0:00:00.645)       0:03:48.024 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********",
                                "Friday 13 March 2026  00:32:11 +0000 (0:00:00.615)       0:03:48.639 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove pam_motd.so rule] ****************************",
                                "Friday 13 March 2026  00:32:12 +0000 (0:00:00.554)       0:03:49.194 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360330.6471221, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360347.723246, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360324.922893, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360319.1326118, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360323.0768747, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360353.9467874, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1773360335.4581244, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy motd file] *************************************",
                                "Friday 13 March 2026  00:32:13 +0000 (0:00:01.015)       0:03:50.209 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue file] ************************************",
                                "Friday 13 March 2026  00:32:14 +0000 (0:00:01.148)       0:03:51.358 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue.net file] ********************************",
                                "Friday 13 March 2026  00:32:15 +0000 (0:00:01.117)       0:03:52.476 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to print the motd] ********************",
                                "Friday 13 March 2026  00:32:16 +0000 (0:00:01.030)       0:03:53.506 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to not print the motd] ****************",
                                "Friday 13 March 2026  00:32:17 +0000 (0:00:00.248)       0:03:53.755 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Include distribution specific install tasks] ********",
                                "Friday 13 March 2026  00:32:17 +0000 (0:00:00.691)       0:03:54.447 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Install rng package] ********************************",
                                "Friday 13 March 2026  00:32:18 +0000 (0:00:00.378)       0:03:54.826 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Remove haveged package] *****************************",
                                "Friday 13 March 2026  00:32:26 +0000 (0:00:08.632)       0:04:03.458 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Manage rng service] *********************************",
                                "Friday 13 March 2026  00:32:28 +0000 (0:00:01.311)       0:04:04.769 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Gather variables for each operating system] ******",
                                "Friday 13 March 2026  00:32:29 +0000 (0:00:01.104)       0:04:05.874 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***",
                                "Friday 13 March 2026  00:32:29 +0000 (0:00:00.276)       0:04:06.150 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***",
                                "Friday 13 March 2026  00:32:29 +0000 (0:00:00.291)       0:04:06.442 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Populate service facts] **************************",
                                "Friday 13 March 2026  00:32:30 +0000 (0:00:00.307)       0:04:06.749 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include distribution specific timer tasks] *******",
                                "Friday 13 March 2026  00:32:34 +0000 (0:00:04.610)       0:04:11.360 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Disable apt-daily timers] ************************",
                                "Friday 13 March 2026  00:32:35 +0000 (0:00:00.398)       0:04:11.759 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include service tasks] ***************************",
                                "Friday 13 March 2026  00:32:35 +0000 (0:00:00.345)       0:04:12.105 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup services] ********************************",
                                "Friday 13 March 2026  00:32:35 +0000 (0:00:00.442)       0:04:12.547 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include packages tasks] **************************",
                                "Friday 13 March 2026  00:32:36 +0000 (0:00:00.339)       0:04:12.887 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup installed packages] **********************",
                                "Friday 13 March 2026  00:32:36 +0000 (0:00:00.382)       0:04:13.269 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloudinit package] ************************",
                                "Friday 13 March 2026  00:33:09 +0000 (0:00:32.796)       0:04:46.066 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********",
                                "Friday 13 March 2026  00:33:18 +0000 (0:00:08.631)       0:04:54.697 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove useless packages from the cache] **********",
                                "Friday 13 March 2026  00:33:26 +0000 (0:00:08.069)       0:05:02.767 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***",
                                "Friday 13 March 2026  00:33:27 +0000 (0:00:01.577)       0:05:04.344 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include cloudinit tasks] *************************",
                                "Friday 13 March 2026  00:33:33 +0000 (0:00:05.897)       0:05:10.242 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******",
                                "Friday 13 March 2026  00:33:33 +0000 (0:00:00.373)       0:05:10.615 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Install tzdata package] *************************",
                                "Friday 13 March 2026  00:33:34 +0000 (0:00:00.624)       0:05:11.239 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Set timezone to UTC] ****************************",
                                "Friday 13 March 2026  00:33:36 +0000 (0:00:01.774)       0:05:13.014 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Create /etc/adjtime file] ***********************",
                                "Friday 13 March 2026  00:33:37 +0000 (0:00:00.669)       0:05:13.684 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************",
                                "Friday 13 March 2026  00:33:37 +0000 (0:00:00.209)       0:05:13.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather variables for each operating system] ******",
                                "Friday 13 March 2026  00:33:37 +0000 (0:00:00.300)       0:05:14.194 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_version variable to default value] ****",
                                "Friday 13 March 2026  00:33:37 +0000 (0:00:00.261)       0:05:14.455 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_cli_version variable to default value] ***",
                                "Friday 13 March 2026  00:33:38 +0000 (0:00:00.228)       0:05:14.684 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker version] ***********************",
                                "Friday 13 March 2026  00:33:38 +0000 (0:00:00.264)       0:05:14.948 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker cli version] *******************",
                                "Friday 13 March 2026  00:33:38 +0000 (0:00:00.227)       0:05:15.175 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include block storage tasks] *********************",
                                "Friday 13 March 2026  00:33:38 +0000 (0:00:00.234)       0:05:15.409 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include zram storage tasks] **********************",
                                "Friday 13 March 2026  00:33:38 +0000 (0:00:00.227)       0:05:15.637 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include docker install tasks] ********************",
                                "Friday 13 March 2026  00:33:39 +0000 (0:00:00.224)       0:05:15.861 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove old architecture-dependent repository] ****",
                                "Friday 13 March 2026  00:33:39 +0000 (0:00:00.443)       0:05:16.305 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather package facts] ****************************",
                                "Friday 13 March 2026  00:33:40 +0000 (0:00:00.774)       0:05:17.079 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Check whether packages are installed that should not be installed] ***",
                                "Friday 13 March 2026  00:33:43 +0000 (0:00:02.938)       0:05:20.017 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install apt-transport-https package] *************",
                                "Friday 13 March 2026  00:33:43 +0000 (0:00:00.573)       0:05:20.591 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository gpg key] **************************",
                                "Friday 13 March 2026  00:33:50 +0000 (0:00:06.564)       0:05:27.155 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository] **********************************",
                                "Friday 13 March 2026  00:33:51 +0000 (0:00:01.010)       0:05:28.166 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Update package cache] ****************************",
                                "Friday 13 March 2026  00:33:59 +0000 (0:00:08.197)       0:05:36.364 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker package version] **********************",
                                "Friday 13 March 2026  00:34:02 +0000 (0:00:03.185)       0:05:39.549 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker-cli package version] ******************",
                                "Friday 13 March 2026  00:34:04 +0000 (0:00:01.200)       0:05:40.750 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unlock containerd package] ***********************",
                                "Friday 13 March 2026  00:34:05 +0000 (0:00:01.452)       0:05:42.202 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install containerd package] **********************",
                                "Friday 13 March 2026  00:34:06 +0000 (0:00:01.004)       0:05:43.207 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Lock containerd package] *************************",
                                "Friday 13 March 2026  00:34:16 +0000 (0:00:09.642)       0:05:52.849 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker-cli package] **********************",
                                "Friday 13 March 2026  00:34:17 +0000 (0:00:00.952)       0:05:53.802 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker package] **************************",
                                "Friday 13 March 2026  00:34:26 +0000 (0:00:09.176)       0:06:02.979 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unblock installation of python docker packages] ***",
                                "Friday 13 March 2026  00:34:37 +0000 (0:00:10.797)       0:06:13.777 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package] ******************",
                                "Friday 13 March 2026  00:34:38 +0000 (0:00:01.224)       0:06:15.001 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package from Debian Sid] ***",
                                "Friday 13 March 2026  00:34:38 +0000 (0:00:00.529)       0:06:15.531 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***",
                                "Friday 13 March 2026  00:34:42 +0000 (0:00:03.866)       0:06:19.397 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***",
                                "Friday 13 March 2026  00:34:43 +0000 (0:00:00.671)       0:06:20.069 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***",
                                "Friday 13 March 2026  00:34:43 +0000 (0:00:00.563)       0:06:20.632 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***",
                                "Friday 13 March 2026  00:34:44 +0000 (0:00:00.510)       0:06:21.143 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install packages required by docker login] *******",
                                "Friday 13 March 2026  00:34:45 +0000 (0:00:00.513)       0:06:21.656 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Ensure that some packages are not installed] *****",
                                "Friday 13 March 2026  00:34:45 +0000 (0:00:00.553)       0:06:22.209 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include config tasks] ****************************",
                                "Friday 13 March 2026  00:34:47 +0000 (0:00:01.825)       0:06:24.035 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create plugins directory] ************************",
                                "Friday 13 March 2026  00:34:48 +0000 (0:00:00.886)       0:06:24.922 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create systemd overlay directory] ****************",
                                "Friday 13 March 2026  00:34:49 +0000 (0:00:00.842)       0:06:25.764 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy systemd overlay file] ***********************",
                                "Friday 13 March 2026  00:34:49 +0000 (0:00:00.865)       0:06:26.629 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***",
                                "Friday 13 March 2026  00:34:51 +0000 (0:00:01.532)       0:06:28.161 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy limits configuration file] ******************",
                                "Friday 13 March 2026  00:34:52 +0000 (0:00:01.216)       0:06:29.378 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy daemon.json configuration file] *************",
                                "Friday 13 March 2026  00:34:54 +0000 (0:00:01.270)       0:06:30.649 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include service tasks] ***************************",
                                "Friday 13 March 2026  00:34:55 +0000 (0:00:01.350)       0:06:31.999 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon] ***************************",
                                "Friday 13 March 2026  00:34:56 +0000 (0:00:00.971)       0:06:32.970 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage service] **********************************",
                                "Friday 13 March 2026  00:34:57 +0000 (0:00:01.382)       0:06:34.353 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage docker socket service] ********************",
                                "Friday 13 March 2026  00:34:58 +0000 (0:00:01.135)       0:06:35.489 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage containerd service] ***********************",
                                "Friday 13 March 2026  00:34:59 +0000 (0:00:01.116)       0:06:36.606 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include bootstrap tasks] *************************",
                                "Friday 13 March 2026  00:35:01 +0000 (0:00:01.262)       0:06:37.868 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.843)       0:06:38.711 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.037)       0:06:38.749 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.043)       0:06:38.793 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.039)       0:06:38.832 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.037)       0:06:38.870 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.044)       0:06:38.915 ********** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.046)       0:06:38.961 ********** ",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Friday 13 March 2026  00:35:02 +0000 (0:00:00.038)       0:06:38.999 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************",
                                "Friday 13 March 2026  00:35:03 +0000 (0:00:01.159)       0:06:40.158 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********",
                                "Friday 13 March 2026  00:35:04 +0000 (0:00:01.459)       0:06:41.618 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Restart docker service] ***************",
                                "Friday 13 March 2026  00:35:06 +0000 (0:00:01.306)       0:06:42.924 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****",
                                "Friday 13 March 2026  00:35:08 +0000 (0:00:02.178)       0:06:45.103 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add user to docker group] ************************",
                                "Friday 13 March 2026  00:35:08 +0000 (0:00:00.081)       0:06:45.184 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Log into private registry and force re-authorization] ***",
                                "Friday 13 March 2026  00:35:09 +0000 (0:00:00.928)       0:06:46.113 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include facts tasks] *****************************",
                                "Friday 13 March 2026  00:35:09 +0000 (0:00:00.522)       0:06:46.635 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create facts directory] **************************",
                                "Friday 13 March 2026  00:35:11 +0000 (0:00:01.111)       0:06:47.747 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy docker fact files] **************************",
                                "Friday 13 March 2026  00:35:11 +0000 (0:00:00.796)       0:06:48.544 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : This install type is not supported] *******",
                                "Friday 13 March 2026  00:35:14 +0000 (0:00:02.451)       0:06:50.995 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Include distribution specific install tasks] ***",
                                "Friday 13 March 2026  00:35:14 +0000 (0:00:00.520)       0:06:51.516 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***",
                                "Friday 13 March 2026  00:35:15 +0000 (0:00:00.762)       0:06:52.279 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******",
                                "Friday 13 March 2026  00:35:16 +0000 (0:00:00.815)       0:06:53.094 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose binary] *************",
                                "Friday 13 March 2026  00:35:17 +0000 (0:00:00.985)       0:06:54.080 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Uninstall docker-compose package] *********",
                                "Friday 13 March 2026  00:35:17 +0000 (0:00:00.493)       0:06:54.573 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose script] ***************",
                                "Friday 13 March 2026  00:35:19 +0000 (0:00:01.483)       0:06:56.056 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****",
                                "Friday 13 March 2026  00:35:19 +0000 (0:00:00.486)       0:06:56.543 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********",
                                "Friday 13 March 2026  00:35:28 +0000 (0:00:08.455)       0:07:04.998 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Enable osism.target] **********************",
                                "Friday 13 March 2026  00:35:29 +0000 (0:00:01.438)       0:07:06.436 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****",
                                "Friday 13 March 2026  00:35:31 +0000 (0:00:01.559)       0:07:07.996 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Friday 13 March 2026  00:35:32 +0000 (0:00:01.508)       0:07:09.504 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Friday 13 March 2026  00:35:33 +0000 (0:00:00.837)       0:07:10.341 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check minimum and maximum number of servers] *****",
                                "Friday 13 March 2026  00:35:34 +0000 (0:00:00.822)       0:07:11.164 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Gather variables for each operating system] ******",
                                "Friday 13 March 2026  00:35:35 +0000 (0:00:00.539)       0:07:11.703 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***",
                                "Friday 13 March 2026  00:35:35 +0000 (0:00:00.551)       0:07:12.255 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_key_file variable to default value] ***",
                                "Friday 13 March 2026  00:35:36 +0000 (0:00:00.756)       0:07:13.012 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Populate service facts] **************************",
                                "Friday 13 March 2026  00:35:36 +0000 (0:00:00.513)       0:07:13.526 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage timesyncd service] ************************",
                                "Friday 13 March 2026  00:35:42 +0000 (0:00:05.378)       0:07:18.904 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Include distribution specific install tasks] *****",
                                "Friday 13 March 2026  00:35:42 +0000 (0:00:00.481)       0:07:19.386 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Install package] *********************************",
                                "Friday 13 March 2026  00:35:43 +0000 (0:00:00.800)       0:07:20.187 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage chrony service] ***************************",
                                "Friday 13 March 2026  00:35:45 +0000 (0:00:01.766)       0:07:21.954 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check if configuration file exists] **************",
                                "Friday 13 March 2026  00:35:47 +0000 (0:00:02.035)       0:07:23.989 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Copy configuration file] *************************",
                                "Friday 13 March 2026  00:35:48 +0000 (0:00:00.811)       0:07:24.801 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Include distribution specific install tasks] ******",
                                "Friday 13 March 2026  00:35:49 +0000 (0:00:01.814)       0:07:26.616 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Install lldpd package] ****************************",
                                "Friday 13 March 2026  00:35:50 +0000 (0:00:00.789)       0:07:27.405 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Manage lldpd service] *****************************",
                                "Friday 13 March 2026  00:36:00 +0000 (0:00:09.257)       0:07:36.662 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********",
                                "Friday 13 March 2026  00:36:02 +0000 (0:00:01.985)       0:07:38.648 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************",
                                "Friday 13 March 2026  00:36:03 +0000 (0:00:01.257)       0:07:39.906 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap role part 2] *********************************************",
                                "",
                                "TASK [Include hardening role] **************************************************",
                                "Friday 13 March 2026  00:36:04 +0000 (0:00:01.177)       0:07:41.084 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 3] ********************************************",
                                "",
                                "TASK [osism.services.journald : Copy configuration file] ***********************",
                                "Friday 13 March 2026  00:36:05 +0000 (0:00:00.722)       0:07:41.806 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.journald : Manage journald service] ***********************",
                                "Friday 13 March 2026  00:36:06 +0000 (0:00:01.413)       0:07:43.220 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Include auditd role] *****************************************************",
                                "Friday 13 March 2026  00:36:07 +0000 (0:00:01.344)       0:07:44.564 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Include smartd role] *****************************************************",
                                "Friday 13 March 2026  00:36:08 +0000 (0:00:00.660)       0:07:45.224 ********** ",
                                "\u001b[0;36mincluded: osism.services.smartd for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Include distribution specific install tasks] *****",
                                "Friday 13 March 2026  00:36:09 +0000 (0:00:00.835)       0:07:46.060 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Install smartmontools package] *******************",
                                "Friday 13 March 2026  00:36:10 +0000 (0:00:00.786)       0:07:46.846 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Create /var/log/smartd directory] ****************",
                                "Friday 13 March 2026  00:36:19 +0000 (0:00:09.117)       0:07:55.963 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Copy smartmontools configuration file] ***********",
                                "Friday 13 March 2026  00:36:20 +0000 (0:00:00.711)       0:07:56.674 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Manage smartd service] ***************************",
                                "Friday 13 March 2026  00:36:21 +0000 (0:00:01.219)       0:07:57.894 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.journald : Restart journald service] ***********",
                                "Friday 13 March 2026  00:36:22 +0000 (0:00:01.715)       0:07:59.609 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************",
                                "Friday 13 March 2026  00:36:24 +0000 (0:00:01.266)       0:08:00.876 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Set state bootstrap] *****************************************************",
                                "",
                                "TASK [Set osism.bootstrap.status fact] *****************************************",
                                "Friday 13 March 2026  00:36:25 +0000 (0:00:01.033)       0:08:01.909 ********** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Friday 13 March 2026  00:36:26 +0000 (0:00:00.911)       0:08:02.821 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Friday 13 March 2026  00:36:26 +0000 (0:00:00.777)       0:08:03.599 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set osism.bootstrap.timestamp fact] **************************************",
                                "Friday 13 March 2026  00:36:28 +0000 (0:00:01.134)       0:08:04.733 ********** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Friday 13 March 2026  00:36:29 +0000 (0:00:00.922)       0:08:05.656 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Friday 13 March 2026  00:36:29 +0000 (0:00:00.829)       0:08:06.485 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=70  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=66  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:36:30 +0000 (0:00:01.039)       0:08:07.525 ********** ",
                                "=============================================================================== ",
                                "osism.commons.packages : Install required packages --------------------- 81.61s",
                                "osism.commons.packages : Download required packages -------------------- 37.34s",
                                "osism.commons.cleanup : Cleanup installed packages --------------------- 32.80s",
                                "osism.commons.repository : Update package cache ------------------------ 18.64s",
                                "osism.commons.systohc : Install util-linux-extra package --------------- 11.36s",
                                "osism.services.docker : Install docker package ------------------------- 10.80s",
                                "osism.commons.packages : Remove dependencies that are no longer required --- 9.95s",
                                "osism.services.docker : Install containerd package ---------------------- 9.64s",
                                "osism.services.lldpd : Install lldpd package ---------------------------- 9.26s",
                                "osism.services.docker : Install docker-cli package ---------------------- 9.18s",
                                "osism.services.smartd : Install smartmontools package ------------------- 9.12s",
                                "osism.services.rng : Install rng package -------------------------------- 8.63s",
                                "osism.commons.cleanup : Remove cloudinit package ------------------------ 8.63s",
                                "osism.commons.docker_compose : Install docker-compose-plugin package ---- 8.46s",
                                "osism.services.docker : Add repository ---------------------------------- 8.20s",
                                "osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.07s",
                                "osism.services.docker : Install apt-transport-https package ------------- 6.56s",
                                "osism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 5.91s",
                                "osism.commons.cleanup : Remove dependencies that are no longer required --- 5.90s",
                                "osism.services.chrony : Populate service facts -------------------------- 5.38s",
                                "+ osism apply fail2ban",
                                "\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m",
                                "\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask dd2ed9d3-1d42-40d1-934e-9f5944ffab6b (fail2ban) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:36:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task dd2ed9d3-1d42-40d1-934e-9f5944ffab6b (fail2ban) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role fail2ban] *****************************************************",
                                "",
                                "TASK [osism.services.fail2ban : Include distribution specific install tasks] ***",
                                "Friday 13 March 2026  00:36:48 +0000 (0:00:00.242)       0:00:00.242 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Install fail2ban package] **********************",
                                "Friday 13 March 2026  00:36:49 +0000 (0:00:01.011)       0:00:01.253 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Copy configuration files] **********************",
                                "Friday 13 March 2026  00:36:59 +0000 (0:00:10.648)       0:00:11.902 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Manage fail2ban service] ***********************",
                                "Friday 13 March 2026  00:37:01 +0000 (0:00:01.341)       0:00:13.243 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Reload fail2ban configuration] *****************",
                                "Friday 13 March 2026  00:37:02 +0000 (0:00:01.335)       0:00:14.578 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:37:03 +0000 (0:00:01.442)       0:00:16.021 ********** ",
                                "=============================================================================== ",
                                "osism.services.fail2ban : Install fail2ban package --------------------- 10.65s",
                                "osism.services.fail2ban : Reload fail2ban configuration ----------------- 1.44s",
                                "osism.services.fail2ban : Copy configuration files ---------------------- 1.34s",
                                "osism.services.fail2ban : Manage fail2ban service ----------------------- 1.33s",
                                "osism.services.fail2ban : Include distribution specific install tasks --- 1.01s",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ osism apply network",
                                "\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m",
                                "\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 43595a2f-1241-4c1c-91aa-86c7a57b59e5 (network) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:37:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 43595a2f-1241-4c1c-91aa-86c7a57b59e5 (network) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role network] ******************************************************",
                                "",
                                "TASK [osism.commons.network : Gather variables for each operating system] ******",
                                "Friday 13 March 2026  00:37:19 +0000 (0:00:00.187)       0:00:00.187 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include type specific tasks] *********************",
                                "Friday 13 March 2026  00:37:20 +0000 (0:00:00.611)       0:00:00.799 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install required packages] ***********************",
                                "Friday 13 March 2026  00:37:21 +0000 (0:00:01.070)       0:00:01.870 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove ifupdown package] *************************",
                                "Friday 13 March 2026  00:37:23 +0000 (0:00:01.958)       0:00:03.829 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create required directories] *********************",
                                "Friday 13 March 2026  00:37:25 +0000 (0:00:01.732)       0:00:05.561 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Prepare netplan configuration template] **********",
                                "Friday 13 March 2026  00:37:26 +0000 (0:00:00.915)       0:00:06.477 ********** ",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy netplan configuration] **********************",
                                "Friday 13 March 2026  00:37:29 +0000 (0:00:03.080)       0:00:09.557 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove netplan configuration template] ***********",
                                "Friday 13 March 2026  00:37:30 +0000 (0:00:01.596)       0:00:11.154 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Check if path for interface file exists] *********",
                                "Friday 13 March 2026  00:37:32 +0000 (0:00:01.649)       0:00:12.803 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy interfaces file] ****************************",
                                "Friday 13 March 2026  00:37:33 +0000 (0:00:01.065)       0:00:13.868 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install package networkd-dispatcher] *************",
                                "Friday 13 March 2026  00:37:34 +0000 (0:00:00.637)       0:00:14.506 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy dispatcher scripts] *************************",
                                "Friday 13 March 2026  00:37:36 +0000 (0:00:02.258)       0:00:16.764 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Manage service networkd-dispatcher] **************",
                                "Friday 13 March 2026  00:37:37 +0000 (0:00:00.761)       0:00:17.526 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include cleanup tasks] ***************************",
                                "Friday 13 March 2026  00:37:38 +0000 (0:00:01.505)       0:00:19.032 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Friday 13 March 2026  00:37:39 +0000 (0:00:01.068)       0:00:20.100 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Set network_configured_files fact] ***************",
                                "Friday 13 March 2026  00:37:40 +0000 (0:00:01.014)       0:00:21.114 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Friday 13 March 2026  00:37:41 +0000 (0:00:00.567)       0:00:21.681 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include dummy interfaces] ************************",
                                "Friday 13 March 2026  00:37:42 +0000 (0:00:01.075)       0:00:22.757 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include vxlan interfaces] ************************",
                                "Friday 13 March 2026  00:37:42 +0000 (0:00:00.660)       0:00:23.417 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-manager, testbed-node-1, testbed-node-0, testbed-node-4, testbed-node-2, testbed-node-3, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd netdev files] ************",
                                "Friday 13 March 2026  00:37:47 +0000 (0:00:04.420)       0:00:27.838 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd network files] ***********",
                                "Friday 13 March 2026  00:37:52 +0000 (0:00:04.944)       0:00:32.783 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include networkd cleanup tasks] ******************",
                                "Friday 13 March 2026  00:37:57 +0000 (0:00:05.025)       0:00:37.808 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Friday 13 March 2026  00:37:58 +0000 (0:00:00.988)       0:00:38.797 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Friday 13 March 2026  00:37:59 +0000 (0:00:01.042)       0:00:39.840 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include network extra init] **********************",
                                "Friday 13 March 2026  00:38:00 +0000 (0:00:00.771)       0:00:40.611 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init script] ****************",
                                "Friday 13 March 2026  00:38:01 +0000 (0:00:01.086)       0:00:41.697 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init systemd service] *******",
                                "Friday 13 March 2026  00:38:01 +0000 (0:00:00.545)       0:00:42.242 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Enable and start network-extra-init service] *****",
                                "Friday 13 March 2026  00:38:02 +0000 (0:00:00.650)       0:00:42.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Disable and stop network-extra-init service] *****",
                                "Friday 13 March 2026  00:38:03 +0000 (0:00:00.582)       0:00:43.476 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init systemd service] *******",
                                "Friday 13 March 2026  00:38:04 +0000 (0:00:01.597)       0:00:45.074 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init script] ****************",
                                "Friday 13 March 2026  00:38:05 +0000 (0:00:00.993)       0:00:46.067 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************",
                                "Friday 13 March 2026  00:38:07 +0000 (0:00:02.363)       0:00:48.431 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********",
                                "Friday 13 March 2026  00:38:08 +0000 (0:00:00.803)       0:00:49.234 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:38:09 +0000 (0:00:00.521)       0:00:49.756 ********** ",
                                "=============================================================================== ",
                                "osism.commons.network : Create systemd networkd network files ----------- 5.03s",
                                "osism.commons.network : Create systemd networkd netdev files ------------ 4.94s",
                                "osism.commons.network : Include vxlan interfaces ------------------------ 4.42s",
                                "osism.commons.network : Prepare netplan configuration template ---------- 3.08s",
                                "osism.commons.network : Remove network-extra-init script ---------------- 2.36s",
                                "osism.commons.network : Install package networkd-dispatcher ------------- 2.26s",
                                "osism.commons.network : Install required packages ----------------------- 1.96s",
                                "osism.commons.network : Remove ifupdown package ------------------------- 1.73s",
                                "osism.commons.network : Remove netplan configuration template ----------- 1.65s",
                                "osism.commons.network : Disable and stop network-extra-init service ----- 1.60s",
                                "osism.commons.network : Copy netplan configuration ---------------------- 1.60s",
                                "osism.commons.network : Manage service networkd-dispatcher -------------- 1.51s",
                                "osism.commons.network : Include network extra init ---------------------- 1.09s",
                                "osism.commons.network : Remove unused configuration files --------------- 1.08s",
                                "osism.commons.network : Include type specific tasks --------------------- 1.07s",
                                "osism.commons.network : Include cleanup tasks --------------------------- 1.07s",
                                "osism.commons.network : Check if path for interface file exists --------- 1.07s",
                                "osism.commons.network : List existing configuration files --------------- 1.04s",
                                "osism.commons.network : List existing configuration files --------------- 1.01s",
                                "osism.commons.network : Remove network-extra-init systemd service ------- 0.99s",
                                "+ osism apply wireguard",
                                "\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m",
                                "\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bf55688f-ea1a-405e-95fa-e6c0ceb11bb2 (wireguard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:38:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bf55688f-ea1a-405e-95fa-e6c0ceb11bb2 (wireguard) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role wireguard] ****************************************************",
                                "",
                                "TASK [osism.services.wireguard : Install iptables package] *********************",
                                "Friday 13 March 2026  00:38:25 +0000 (0:00:00.168)       0:00:00.168 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Install wireguard package] ********************",
                                "Friday 13 March 2026  00:38:27 +0000 (0:00:01.144)       0:00:01.312 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create public and private key - server] *******",
                                "Friday 13 March 2026  00:38:32 +0000 (0:00:05.011)       0:00:06.324 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create preshared key] *************************",
                                "Friday 13 March 2026  00:38:32 +0000 (0:00:00.498)       0:00:06.823 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get preshared key] ****************************",
                                "Friday 13 March 2026  00:38:32 +0000 (0:00:00.375)       0:00:07.199 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get public key - server] **********************",
                                "Friday 13 March 2026  00:38:33 +0000 (0:00:00.558)       0:00:07.758 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get private key - server] *********************",
                                "Friday 13 March 2026  00:38:33 +0000 (0:00:00.383)       0:00:08.142 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy wg0.conf configuration file] *************",
                                "Friday 13 March 2026  00:38:34 +0000 (0:00:00.378)       0:00:08.520 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy client configuration files] **************",
                                "Friday 13 March 2026  00:38:35 +0000 (0:00:01.133)       0:00:09.654 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********",
                                "Friday 13 March 2026  00:38:36 +0000 (0:00:00.893)       0:00:10.547 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************",
                                "Friday 13 March 2026  00:38:37 +0000 (0:00:01.661)       0:00:12.209 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:38:38 +0000 (0:00:00.889)       0:00:13.098 ********** ",
                                "=============================================================================== ",
                                "osism.services.wireguard : Install wireguard package -------------------- 5.01s",
                                "osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.66s",
                                "osism.services.wireguard : Install iptables package --------------------- 1.14s",
                                "osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.13s",
                                "osism.services.wireguard : Copy client configuration files -------------- 0.89s",
                                "osism.services.wireguard : Restart wg0 service -------------------------- 0.89s",
                                "osism.services.wireguard : Get preshared key ---------------------------- 0.56s",
                                "osism.services.wireguard : Create public and private key - server ------- 0.50s",
                                "osism.services.wireguard : Get public key - server ---------------------- 0.38s",
                                "osism.services.wireguard : Get private key - server --------------------- 0.38s",
                                "osism.services.wireguard : Create preshared key ------------------------- 0.38s",
                                "+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0",
                                "100    14  100    14    0     0    197      0 --:--:-- --:--:-- --:--:--   200",
                                "+ osism apply --environment custom workarounds",
                                "\u001b[32m2026-03-13 00:38:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m",
                                "\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m",
                                "\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f20912d7-2bba-457a-8259-9c7be1649d3c (workarounds) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:38:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f20912d7-2bba-457a-8259-9c7be1649d3c (workarounds) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on virtualization_role] ********************************",
                                "Friday 13 March 2026  00:38:55 +0000 (0:00:00.112)       0:00:00.112 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on the manager node] *************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Friday 13 March 2026  00:38:56 +0000 (0:00:00.669)       0:00:00.782 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on all other nodes] **************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Friday 13 March 2026  00:38:58 +0000 (0:00:02.086)       0:00:02.868 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Add custom CA certificates to non-manager nodes] *************************",
                                "",
                                "TASK [Copy custom CA certificates] *********************************************",
                                "Friday 13 March 2026  00:39:00 +0000 (0:00:01.847)       0:00:04.716 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "",
                                "TASK [Run update-ca-certificates] **********************************************",
                                "Friday 13 March 2026  00:39:01 +0000 (0:00:01.422)       0:00:06.138 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Run update-ca-trust] *****************************************************",
                                "Friday 13 March 2026  00:39:05 +0000 (0:00:03.509)       0:00:09.647 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Add a workaround service] ************************************************",
                                "",
                                "TASK [Copy workarounds.sh scripts] *********************************************",
                                "Friday 13 March 2026  00:39:05 +0000 (0:00:00.610)       0:00:10.258 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Copy workarounds systemd unit file] **************************************",
                                "Friday 13 March 2026  00:39:07 +0000 (0:00:01.790)       0:00:12.049 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reload systemd daemon] ***************************************************",
                                "Friday 13 March 2026  00:39:09 +0000 (0:00:01.590)       0:00:13.639 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Enable workarounds.service (Debian)] *************************************",
                                "Friday 13 March 2026  00:39:10 +0000 (0:00:01.486)       0:00:15.126 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Enable and start workarounds.service (RedHat)] ***************************",
                                "Friday 13 March 2026  00:39:12 +0000 (0:00:01.813)       0:00:16.940 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************",
                                "",
                                "TASK [Install python3-docker] **************************************************",
                                "Friday 13 March 2026  00:39:13 +0000 (0:00:00.649)       0:00:17.589 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:39:15 +0000 (0:00:02.843)       0:00:20.432 ********** ",
                                "=============================================================================== ",
                                "Run update-ca-certificates ---------------------------------------------- 3.51s",
                                "Install python3-docker -------------------------------------------------- 2.84s",
                                "Apply netplan configuration --------------------------------------------- 2.09s",
                                "Apply netplan configuration --------------------------------------------- 1.85s",
                                "Enable workarounds.service (Debian) ------------------------------------- 1.81s",
                                "Copy workarounds.sh scripts --------------------------------------------- 1.79s",
                                "Copy workarounds systemd unit file -------------------------------------- 1.59s",
                                "Reload systemd daemon --------------------------------------------------- 1.49s",
                                "Copy custom CA certificates --------------------------------------------- 1.42s",
                                "Group hosts based on virtualization_role -------------------------------- 0.67s",
                                "Enable and start workarounds.service (RedHat) --------------------------- 0.65s",
                                "Run update-ca-trust ----------------------------------------------------- 0.61s",
                                "+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m",
                                "\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4c7df022-a09d-40d7-84a9-9061b406155a (reboot) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:39:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4c7df022-a09d-40d7-84a9-9061b406155a (reboot) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:32 +0000 (0:00:00.149)       0:00:00.149 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:32 +0000 (0:00:00.096)       0:00:00.246 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:33 +0000 (0:00:00.898)       0:00:01.144 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:33 +0000 (0:00:00.095)       0:00:01.240 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:33 +0000 (0:00:00.085)       0:00:01.325 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:33 +0000 (0:00:00.630)       0:00:01.955 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:34 +0000 (0:00:00.105)       0:00:02.061 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:34 +0000 (0:00:00.154)       0:00:02.216 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:34 +0000 (0:00:00.667)       0:00:02.884 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:34 +0000 (0:00:00.104)       0:00:02.988 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:35 +0000 (0:00:00.094)       0:00:03.083 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:35 +0000 (0:00:00.653)       0:00:03.736 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:35 +0000 (0:00:00.101)       0:00:03.837 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:35 +0000 (0:00:00.096)       0:00:03.934 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:36 +0000 (0:00:00.685)       0:00:04.619 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Friday 13 March 2026  00:39:36 +0000 (0:00:00.125)       0:00:04.745 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Friday 13 March 2026  00:39:36 +0000 (0:00:00.109)       0:00:04.854 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Friday 13 March 2026  00:39:37 +0000 (0:00:00.714)       0:00:05.569 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:39:37 +0000 (0:00:00.033)       0:00:05.602 ********** ",
                                "=============================================================================== ",
                                "Reboot system - do not wait for the reboot to complete ------------------ 4.25s",
                                "Exit playbook, if user did not mean to reboot systems ------------------- 0.64s",
                                "Reboot system - wait for the reboot to complete ------------------------- 0.57s",
                                "+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-03-13 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m",
                                "\u001b[32m2026-03-13 00:39:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 899bfc0f-3626-4da4-91f0-7044c9d66a42 (wait-for-connection) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:39:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 899bfc0f-3626-4da4-91f0-7044c9d66a42 (wait-for-connection) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wait until remote systems are reachable] *********************************",
                                "",
                                "TASK [Wait until remote system is reachable] ***********************************",
                                "Friday 13 March 2026  00:39:53 +0000 (0:00:00.243)       0:00:00.243 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:40:05 +0000 (0:00:11.488)       0:00:11.732 ********** ",
                                "=============================================================================== ",
                                "Wait until remote system is reachable ---------------------------------- 11.49s",
                                "+ osism apply hddtemp",
                                "\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m",
                                "\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca56fc81-55da-4470-a81d-f25cc4eddfbf (hddtemp) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:40:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ca56fc81-55da-4470-a81d-f25cc4eddfbf (hddtemp) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role hddtemp] ******************************************************",
                                "",
                                "TASK [osism.services.hddtemp : Gather variables for each operating system] *****",
                                "Friday 13 March 2026  00:40:22 +0000 (0:00:00.233)       0:00:00.233 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific install tasks] ****",
                                "Friday 13 March 2026  00:40:23 +0000 (0:00:00.625)       0:00:00.859 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Remove hddtemp package] *************************",
                                "Friday 13 March 2026  00:40:24 +0000 (0:00:01.005)       0:00:01.864 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************",
                                "Friday 13 March 2026  00:40:25 +0000 (0:00:01.874)       0:00:03.739 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Check if drivetemp module is available] *********",
                                "Friday 13 March 2026  00:40:26 +0000 (0:00:01.027)       0:00:04.766 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************",
                                "Friday 13 March 2026  00:40:28 +0000 (0:00:01.192)       0:00:05.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Install lm-sensors] *****************************",
                                "Friday 13 March 2026  00:40:28 +0000 (0:00:00.772)       0:00:06.731 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific service tasks] ****",
                                "Friday 13 March 2026  00:40:42 +0000 (0:00:13.955)       0:00:20.687 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Manage lm-sensors service] **********************",
                                "Friday 13 March 2026  00:40:44 +0000 (0:00:01.118)       0:00:21.806 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:40:45 +0000 (0:00:01.769)       0:00:23.575 ********** ",
                                "=============================================================================== ",
                                "osism.services.hddtemp : Install lm-sensors ---------------------------- 13.96s",
                                "osism.services.hddtemp : Remove hddtemp package ------------------------- 1.87s",
                                "osism.services.hddtemp : Manage lm-sensors service ---------------------- 1.77s",
                                "osism.services.hddtemp : Check if drivetemp module is available --------- 1.19s",
                                "osism.services.hddtemp : Include distribution specific service tasks ---- 1.12s",
                                "osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.03s",
                                "osism.services.hddtemp : Include distribution specific install tasks ---- 1.01s",
                                "osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.77s",
                                "osism.services.hddtemp : Gather variables for each operating system ----- 0.63s",
                                "++ semver latest 7.1.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sudo systemctl restart manager.service",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "ARA in ceph-ansible already disabled.",
                                "ARA in kolla-ansible already disabled.",
                                "ARA in osism-ansible already disabled.",
                                "ARA in osism-kubernetes already disabled.",
                                "+ osism apply gather-facts",
                                "\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5f6f437b-413b-4f54-9e59-ce9e6a25bddd (gather-facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5f6f437b-413b-4f54-9e59-ce9e6a25bddd (gather-facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:42:21 +0000 (0:00:00.160)       0:00:00.160 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Friday 13 March 2026  00:42:30 +0000 (0:00:08.252)       0:00:08.412 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:42:30 +0000 (0:00:00.447)       0:00:08.859 ********** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 8.25s",
                                "Gather facts for all hosts ---------------------------------------------- 0.45s",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry",
                                "+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images",
                                "+ [[ false == \\t\\r\\u\\e ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001a-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:42:31.486413Z",
                            "start": "2026-03-13T00:19:06.441301Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001a",
                        "name": "Deploy manager + bootstrap nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "not _nutshell | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T00:42:32.119549Z",
                            "start": "2026-03-13T00:42:31.609427Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001b",
                        "name": "Deploy services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "/opt/configuration/scripts/deploy-in-a-nutshell.sh"
                            ],
                            "delta": "0:33:53.899074",
                            "end": "2026-03-13 01:16:26.343380",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 /opt/configuration/scripts/deploy-in-a-nutshell.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001c-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 00:42:32.444306",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n\n# PULL IMAGES\n\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# PULL IMAGES'\n+ echo\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-03-13 00:42:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m\n\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77abe84d-01e8-47ab-a1c1-42460e02917d (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77abe84d-01e8-47ab-a1c1-42460e02917d is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-03-13 00:42:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m\n\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m\n\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask aff5573c-3d9a-4d83-bd2d-d50ad7552953 (wipe-partitions) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task aff5573c-3d9a-4d83-bd2d-d50ad7552953 (wipe-partitions) has been started and output is visible here.\u001b[0m\n\nPLAY [Wipe partitions] *********************************************************\n\nTASK [Find all logical devices owned by UID 167] *******************************\nFriday 13 March 2026  00:43:00 +0000 (0:00:00.095)       0:00:00.095 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Remove all rook related logical devices] *********************************\nFriday 13 March 2026  00:43:00 +0000 (0:00:00.534)       0:00:00.629 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Find all logical devices with prefix ceph] *******************************\nFriday 13 March 2026  00:43:01 +0000 (0:00:00.369)       0:00:00.999 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Remove all ceph related logical devices] *********************************\nFriday 13 March 2026  00:43:01 +0000 (0:00:00.684)       0:00:01.684 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check device availability] ***********************************************\nFriday 13 March 2026  00:43:02 +0000 (0:00:00.261)       0:00:01.945 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Wipe partitions with wipefs] *********************************************\nFriday 13 March 2026  00:43:03 +0000 (0:00:01.230)       0:00:03.176 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Overwrite first 32M with zeros] ******************************************\nFriday 13 March 2026  00:43:04 +0000 (0:00:01.570)       0:00:04.746 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Reload udev rules] *******************************************************\nFriday 13 March 2026  00:43:07 +0000 (0:00:02.237)       0:00:06.984 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Request device events from the kernel] ***********************************\nFriday 13 March 2026  00:43:07 +0000 (0:00:00.554)       0:00:07.538 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:43:08 +0000 (0:00:00.648)       0:00:08.186 ********** \n=============================================================================== \nOverwrite first 32M with zeros ------------------------------------------ 2.24s\nWipe partitions with wipefs --------------------------------------------- 1.57s\nCheck device availability ----------------------------------------------- 1.23s\nFind all logical devices with prefix ceph ------------------------------- 0.68s\nRequest device events from the kernel ----------------------------------- 0.65s\nReload udev rules ------------------------------------------------------- 0.55s\nFind all logical devices owned by UID 167 ------------------------------- 0.53s\nRemove all rook related logical devices --------------------------------- 0.37s\nRemove all ceph related logical devices --------------------------------- 0.26s\n\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc658442-32df-4315-b988-2f6b632e5043 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cc658442-32df-4315-b988-2f6b632e5043 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nFriday 13 March 2026  00:43:24 +0000 (0:00:00.217)       0:00:00.217 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nFriday 13 March 2026  00:43:25 +0000 (0:00:00.906)       0:00:01.123 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:43:26 +0000 (0:00:01.054)       0:00:02.177 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nFriday 13 March 2026  00:43:32 +0000 (0:00:05.948)       0:00:08.126 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:43:32 +0000 (0:00:00.434)       0:00:08.561 ********** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.95s\nosism.commons.facts : Copy fact files ----------------------------------- 1.05s\nosism.commons.facts : Create custom facts directory --------------------- 0.91s\nGather facts for all hosts ---------------------------------------------- 0.43s\n\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m\n\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8129fd5d-59bf-429f-8120-1b55b222bbf7 (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8129fd5d-59bf-429f-8120-1b55b222bbf7 (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:43:38 +0000 (0:00:00.246)       0:00:00.246 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:43:38 +0000 (0:00:00.213)       0:00:00.460 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:38 +0000 (0:00:00.204)       0:00:00.665 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:39 +0000 (0:00:00.453)       0:00:01.118 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:39 +0000 (0:00:00.170)       0:00:01.289 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:39 +0000 (0:00:00.161)       0:00:01.451 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:39 +0000 (0:00:00.199)       0:00:01.650 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:39 +0000 (0:00:00.178)       0:00:01.829 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:40 +0000 (0:00:00.174)       0:00:02.003 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:40 +0000 (0:00:00.167)       0:00:02.170 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:40 +0000 (0:00:00.186)       0:00:02.357 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:40 +0000 (0:00:00.182)       0:00:02.540 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:40 +0000 (0:00:00.353)       0:00:02.894 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:41 +0000 (0:00:00.501)       0:00:03.395 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:42 +0000 (0:00:00.536)       0:00:03.931 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:42 +0000 (0:00:00.704)       0:00:04.636 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:43 +0000 (0:00:00.295)       0:00:04.931 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:43 +0000 (0:00:00.336)       0:00:05.267 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:43 +0000 (0:00:00.182)       0:00:05.450 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:43 +0000 (0:00:00.183)       0:00:05.633 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:43 +0000 (0:00:00.177)       0:00:05.811 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:44 +0000 (0:00:00.169)       0:00:05.980 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:44 +0000 (0:00:00.185)       0:00:06.165 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:44 +0000 (0:00:00.180)       0:00:06.346 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:44 +0000 (0:00:00.172)       0:00:06.518 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:44 +0000 (0:00:00.186)       0:00:06.704 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:45 +0000 (0:00:00.777)       0:00:07.483 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:45 +0000 (0:00:00.170)       0:00:07.653 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:45 +0000 (0:00:00.172)       0:00:07.826 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.169)       0:00:07.995 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.193)       0:00:08.189 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.149)       0:00:08.339 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.109)       0:00:08.448 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.115)       0:00:08.563 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.115)       0:00:08.679 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nFriday 13 March 2026  00:43:46 +0000 (0:00:00.123)       0:00:08.802 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.150)       0:00:08.953 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.139)       0:00:09.092 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.272)       0:00:09.365 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.161)       0:00:09.526 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.116)       0:00:09.643 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.115)       0:00:09.758 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nFriday 13 March 2026  00:43:47 +0000 (0:00:00.110)       0:00:09.869 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.131)       0:00:10.001 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.123)       0:00:10.124 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.129)       0:00:10.254 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.123)       0:00:10.377 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.122)       0:00:10.500 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nFriday 13 March 2026  00:43:48 +0000 (0:00:00.126)       0:00:10.626 ********** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-34f0c902-d036-5088-bfac-53928dad7c62\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nFriday 13 March 2026  00:43:49 +0000 (0:00:00.327)       0:00:10.953 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:43:50 +0000 (0:00:01.604)       0:00:12.558 ********** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:43:50 +0000 (0:00:00.232)       0:00:12.791 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:51 +0000 (0:00:00.210)       0:00:13.001 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:51 +0000 (0:00:00.365)       0:00:13.367 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:51 +0000 (0:00:00.192)       0:00:13.559 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:51 +0000 (0:00:00.179)       0:00:13.739 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:51 +0000 (0:00:00.174)       0:00:13.913 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:52 +0000 (0:00:00.183)       0:00:14.097 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:52 +0000 (0:00:00.472)       0:00:14.569 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:52 +0000 (0:00:00.188)       0:00:14.757 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:53 +0000 (0:00:00.176)       0:00:14.934 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:53 +0000 (0:00:00.193)       0:00:15.127 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:53 +0000 (0:00:00.379)       0:00:15.507 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:53 +0000 (0:00:00.384)       0:00:15.891 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:54 +0000 (0:00:00.411)       0:00:16.303 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:43:54 +0000 (0:00:00.376)       0:00:16.679 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:55 +0000 (0:00:00.299)       0:00:16.979 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:55 +0000 (0:00:00.346)       0:00:17.325 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:55 +0000 (0:00:00.488)       0:00:17.814 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.179)       0:00:17.993 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.184)       0:00:18.178 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.178)       0:00:18.357 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.188)       0:00:18.545 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.179)       0:00:18.724 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:56 +0000 (0:00:00.175)       0:00:18.900 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:57 +0000 (0:00:00.183)       0:00:19.083 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:57 +0000 (0:00:00.715)       0:00:19.798 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:58 +0000 (0:00:00.196)       0:00:19.995 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:58 +0000 (0:00:00.173)       0:00:20.168 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:43:58 +0000 (0:00:00.187)       0:00:20.355 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nFriday 13 March 2026  00:43:58 +0000 (0:00:00.540)       0:00:20.896 ********** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.186)       0:00:21.082 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.112)       0:00:21.195 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.116)       0:00:21.312 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.116)       0:00:21.428 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.114)       0:00:21.543 ********** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.143)       0:00:21.686 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nFriday 13 March 2026  00:43:59 +0000 (0:00:00.134)       0:00:21.820 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.138)       0:00:21.959 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.123)       0:00:22.082 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.107)       0:00:22.189 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.112)       0:00:22.301 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.253)       0:00:22.555 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.134)       0:00:22.689 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nFriday 13 March 2026  00:44:00 +0000 (0:00:00.152)       0:00:22.842 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nFriday 13 March 2026  00:44:01 +0000 (0:00:00.259)       0:00:23.101 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nFriday 13 March 2026  00:44:01 +0000 (0:00:00.099)       0:00:23.201 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nFriday 13 March 2026  00:44:01 +0000 (0:00:00.120)       0:00:23.322 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nFriday 13 March 2026  00:44:01 +0000 (0:00:00.125)       0:00:23.447 ********** \n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-14d46521-cd63-5c40-b949-080247a26419\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-27536765-3469-5fa1-a30d-2781c44211ac\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nFriday 13 March 2026  00:44:01 +0000 (0:00:00.234)       0:00:23.682 ********** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:44:02 +0000 (0:00:01.076)       0:00:24.758 ********** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:44:03 +0000 (0:00:00.558)       0:00:25.317 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:03 +0000 (0:00:00.284)       0:00:25.601 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.346)       0:00:25.947 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.165)       0:00:26.113 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.184)       0:00:26.297 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.166)       0:00:26.464 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.167)       0:00:26.631 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:04 +0000 (0:00:00.178)       0:00:26.810 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:05 +0000 (0:00:00.180)       0:00:26.990 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:05 +0000 (0:00:00.153)       0:00:27.144 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:05 +0000 (0:00:00.141)       0:00:27.285 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:05 +0000 (0:00:00.607)       0:00:27.893 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:06 +0000 (0:00:00.358)       0:00:28.251 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:06 +0000 (0:00:00.339)       0:00:28.591 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:44:07 +0000 (0:00:00.337)       0:00:28.928 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:07 +0000 (0:00:00.255)       0:00:29.184 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:07 +0000 (0:00:00.311)       0:00:29.496 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:07 +0000 (0:00:00.186)       0:00:29.683 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:07 +0000 (0:00:00.177)       0:00:29.860 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:08 +0000 (0:00:00.207)       0:00:30.067 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:08 +0000 (0:00:00.164)       0:00:30.231 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:08 +0000 (0:00:00.179)       0:00:30.411 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:08 +0000 (0:00:00.486)       0:00:30.897 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:09 +0000 (0:00:00.198)       0:00:31.096 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:09 +0000 (0:00:00.192)       0:00:31.288 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.664)       0:00:31.952 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.207)       0:00:32.160 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.203)       0:00:32.364 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.189)       0:00:32.553 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.192)       0:00:32.746 ********** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nFriday 13 March 2026  00:44:10 +0000 (0:00:00.161)       0:00:32.907 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nFriday 13 March 2026  00:44:11 +0000 (0:00:00.108)       0:00:33.016 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nFriday 13 March 2026  00:44:11 +0000 (0:00:00.185)       0:00:33.201 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nFriday 13 March 2026  00:44:11 +0000 (0:00:00.423)       0:00:33.625 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nFriday 13 March 2026  00:44:11 +0000 (0:00:00.149)       0:00:33.775 ********** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.202)       0:00:33.977 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.181)       0:00:34.159 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.196)       0:00:34.355 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.148)       0:00:34.504 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.134)       0:00:34.638 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nFriday 13 March 2026  00:44:12 +0000 (0:00:00.168)       0:00:34.806 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nFriday 13 March 2026  00:44:13 +0000 (0:00:00.175)       0:00:34.982 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nFriday 13 March 2026  00:44:13 +0000 (0:00:00.159)       0:00:35.141 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nFriday 13 March 2026  00:44:13 +0000 (0:00:00.128)       0:00:35.269 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nFriday 13 March 2026  00:44:13 +0000 (0:00:00.150)       0:00:35.419 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nFriday 13 March 2026  00:44:13 +0000 (0:00:00.180)       0:00:35.600 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nFriday 13 March 2026  00:44:14 +0000 (0:00:00.361)       0:00:35.961 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nFriday 13 March 2026  00:44:14 +0000 (0:00:00.143)       0:00:36.105 ********** \n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-a323501f-4860-533e-be9d-aa017f55891f\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-02d8354f-c2be-5171-944a-5a76777090d4\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nFriday 13 March 2026  00:44:14 +0000 (0:00:00.237)       0:00:36.342 ********** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:44:15 +0000 (0:00:00.865)       0:00:37.208 ********** \n=============================================================================== \nWrite configuration file ------------------------------------------------ 3.55s\nAdd known links to the list of available block devices ------------------ 1.17s\nGet extra vars for Ceph configuration ----------------------------------- 1.00s\nAdd known partitions to the list of available block devices ------------- 0.99s\nPrint configuration data ------------------------------------------------ 0.80s\nAdd known partitions to the list of available block devices ------------- 0.78s\nAdd known partitions to the list of available block devices ------------- 0.72s\nAdd known links to the list of available block devices ------------------ 0.70s\nGet initial list of available block devices ----------------------------- 0.70s\nAdd known partitions to the list of available block devices ------------- 0.66s\nGenerate shared DB/WAL VG names ----------------------------------------- 0.65s\nAdd known links to the list of available block devices ------------------ 0.61s\nGenerate lvm_volumes structure (block + wal) ---------------------------- 0.61s\nPrint DB devices -------------------------------------------------------- 0.61s\nAdd known partitions to the list of available block devices ------------- 0.54s\nSet DB devices config data ---------------------------------------------- 0.54s\nPrint ceph_osd_devices -------------------------------------------------- 0.54s\nAdd known links to the list of available block devices ------------------ 0.54s\nAdd known links to the list of available block devices ------------------ 0.50s\nSet UUIDs for OSD VGs/LVs ----------------------------------------------- 0.50s\n\u001b[32m2026-03-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5aa282cd-b8dc-4f2c-8a92-62ef961101c0 (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-03-13 00:44:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-03-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-03-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-03-13 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-03-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n[master f8e2d0f] 2026-03-13-00-45\n 1 file changed, 30 insertions(+), 9 deletions(-)\n\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m\n\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b90afbba-b381-46bc-9cad-41547468fc50 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b90afbba-b381-46bc-9cad-41547468fc50 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:45:10 +0000 (0:00:00.325)       0:00:00.325 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:45:10 +0000 (0:00:00.288)       0:00:00.614 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:11 +0000 (0:00:00.407)       0:00:01.022 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:11 +0000 (0:00:00.529)       0:00:01.551 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:11 +0000 (0:00:00.183)       0:00:01.734 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:12 +0000 (0:00:00.196)       0:00:01.931 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:12 +0000 (0:00:00.201)       0:00:02.133 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:12 +0000 (0:00:00.206)       0:00:02.339 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:12 +0000 (0:00:00.194)       0:00:02.533 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:12 +0000 (0:00:00.211)       0:00:02.745 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:13 +0000 (0:00:00.209)       0:00:02.955 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:13 +0000 (0:00:00.217)       0:00:03.173 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:13 +0000 (0:00:00.454)       0:00:03.627 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:14 +0000 (0:00:00.597)       0:00:04.225 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:15 +0000 (0:00:00.770)       0:00:04.995 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:16 +0000 (0:00:01.153)       0:00:06.149 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:16 +0000 (0:00:00.336)       0:00:06.485 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:16 +0000 (0:00:00.442)       0:00:06.928 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:17 +0000 (0:00:00.192)       0:00:07.120 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:17 +0000 (0:00:00.215)       0:00:07.336 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:17 +0000 (0:00:00.193)       0:00:07.529 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:17 +0000 (0:00:00.170)       0:00:07.700 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:17 +0000 (0:00:00.164)       0:00:07.864 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:18 +0000 (0:00:00.177)       0:00:08.042 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:18 +0000 (0:00:00.175)       0:00:08.217 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:18 +0000 (0:00:00.168)       0:00:08.386 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:19 +0000 (0:00:00.966)       0:00:09.353 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:19 +0000 (0:00:00.151)       0:00:09.504 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:19 +0000 (0:00:00.187)       0:00:09.692 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:19 +0000 (0:00:00.187)       0:00:09.879 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nFriday 13 March 2026  00:45:20 +0000 (0:00:00.139)       0:00:10.019 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nFriday 13 March 2026  00:45:20 +0000 (0:00:00.098)       0:00:10.117 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nFriday 13 March 2026  00:45:20 +0000 (0:00:00.154)       0:00:10.272 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nFriday 13 March 2026  00:45:22 +0000 (0:00:01.758)       0:00:12.031 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nFriday 13 March 2026  00:45:22 +0000 (0:00:00.172)       0:00:12.204 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nFriday 13 March 2026  00:45:23 +0000 (0:00:01.423)       0:00:13.627 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nFriday 13 March 2026  00:45:23 +0000 (0:00:00.131)       0:00:13.758 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nFriday 13 March 2026  00:45:23 +0000 (0:00:00.121)       0:00:13.880 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.267)       0:00:14.147 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.129)       0:00:14.277 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.135)       0:00:14.413 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.120)       0:00:14.534 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.132)       0:00:14.666 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.128)       0:00:14.795 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nFriday 13 March 2026  00:45:24 +0000 (0:00:00.133)       0:00:14.928 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.132)       0:00:15.060 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.136)       0:00:15.197 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.132)       0:00:15.329 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.119)       0:00:15.449 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.123)       0:00:15.572 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nFriday 13 March 2026  00:45:25 +0000 (0:00:00.241)       0:00:15.814 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nFriday 13 March 2026  00:45:26 +0000 (0:00:00.127)       0:00:15.942 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nFriday 13 March 2026  00:45:26 +0000 (0:00:00.125)       0:00:16.068 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nFriday 13 March 2026  00:45:26 +0000 (0:00:00.617)       0:00:16.685 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nFriday 13 March 2026  00:45:27 +0000 (0:00:00.515)       0:00:17.200 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nFriday 13 March 2026  00:45:27 +0000 (0:00:00.494)       0:00:17.695 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nFriday 13 March 2026  00:45:27 +0000 (0:00:00.130)       0:00:17.826 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nFriday 13 March 2026  00:45:27 +0000 (0:00:00.095)       0:00:17.921 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.098)       0:00:18.019 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.130)       0:00:18.150 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.124)       0:00:18.274 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.114)       0:00:18.389 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.223)       0:00:18.612 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.137)       0:00:18.750 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nFriday 13 March 2026  00:45:28 +0000 (0:00:00.135)       0:00:18.885 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.109)       0:00:18.995 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.124)       0:00:19.120 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.138)       0:00:19.258 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.151)       0:00:19.409 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.151)       0:00:19.561 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.135)       0:00:19.696 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nFriday 13 March 2026  00:45:29 +0000 (0:00:00.137)       0:00:19.833 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.137)       0:00:19.971 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.128)       0:00:20.099 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.292)       0:00:20.392 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.125)       0:00:20.517 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.129)       0:00:20.647 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nFriday 13 March 2026  00:45:30 +0000 (0:00:00.154)       0:00:20.802 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nFriday 13 March 2026  00:45:31 +0000 (0:00:00.142)       0:00:20.944 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nFriday 13 March 2026  00:45:31 +0000 (0:00:00.140)       0:00:21.085 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nFriday 13 March 2026  00:45:31 +0000 (0:00:00.139)       0:00:21.225 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nFriday 13 March 2026  00:45:31 +0000 (0:00:00.148)       0:00:21.373 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nFriday 13 March 2026  00:45:31 +0000 (0:00:00.493)       0:00:21.867 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nFriday 13 March 2026  00:45:32 +0000 (0:00:00.501)       0:00:22.368 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nFriday 13 March 2026  00:45:32 +0000 (0:00:00.153)       0:00:22.522 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'vg_name': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'vg_name': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nFriday 13 March 2026  00:45:32 +0000 (0:00:00.171)       0:00:22.694 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nFriday 13 March 2026  00:45:33 +0000 (0:00:00.306)       0:00:23.000 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nFriday 13 March 2026  00:45:33 +0000 (0:00:00.143)       0:00:23.144 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nFriday 13 March 2026  00:45:33 +0000 (0:00:00.154)       0:00:23.298 ********** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-34f0c902-d036-5088-bfac-53928dad7c62\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:45:33 +0000 (0:00:00.293)       0:00:23.592 ********** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:45:33 +0000 (0:00:00.222)       0:00:23.814 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:34 +0000 (0:00:00.190)       0:00:24.004 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:34 +0000 (0:00:00.370)       0:00:24.375 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:34 +0000 (0:00:00.206)       0:00:24.581 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:34 +0000 (0:00:00.188)       0:00:24.769 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:35 +0000 (0:00:00.490)       0:00:25.259 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:35 +0000 (0:00:00.195)       0:00:25.455 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:35 +0000 (0:00:00.172)       0:00:25.627 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:35 +0000 (0:00:00.213)       0:00:25.841 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:36 +0000 (0:00:00.212)       0:00:26.054 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:36 +0000 (0:00:00.202)       0:00:26.256 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:36 +0000 (0:00:00.418)       0:00:26.675 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:37 +0000 (0:00:00.392)       0:00:27.068 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:37 +0000 (0:00:00.398)       0:00:27.467 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:38 +0000 (0:00:00.499)       0:00:27.966 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:38 +0000 (0:00:00.499)       0:00:28.465 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:39 +0000 (0:00:00.513)       0:00:28.979 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:39 +0000 (0:00:00.222)       0:00:29.201 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:39 +0000 (0:00:00.183)       0:00:29.384 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:39 +0000 (0:00:00.188)       0:00:29.573 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:39 +0000 (0:00:00.197)       0:00:29.771 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:40 +0000 (0:00:00.211)       0:00:29.983 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:40 +0000 (0:00:00.185)       0:00:30.168 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:40 +0000 (0:00:00.213)       0:00:30.382 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:40 +0000 (0:00:00.212)       0:00:30.594 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:41 +0000 (0:00:00.766)       0:00:31.361 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:41 +0000 (0:00:00.200)       0:00:31.561 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:42 +0000 (0:00:00.664)       0:00:32.225 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:45:42 +0000 (0:00:00.186)       0:00:32.411 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nFriday 13 March 2026  00:45:42 +0000 (0:00:00.202)       0:00:32.614 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nFriday 13 March 2026  00:45:42 +0000 (0:00:00.134)       0:00:32.749 ********** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nFriday 13 March 2026  00:45:43 +0000 (0:00:00.217)       0:00:32.967 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nFriday 13 March 2026  00:45:44 +0000 (0:00:01.858)       0:00:34.826 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nFriday 13 March 2026  00:45:45 +0000 (0:00:00.139)       0:00:34.966 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nFriday 13 March 2026  00:45:46 +0000 (0:00:01.377)       0:00:36.343 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nFriday 13 March 2026  00:45:46 +0000 (0:00:00.137)       0:00:36.481 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nFriday 13 March 2026  00:45:46 +0000 (0:00:00.163)       0:00:36.645 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nFriday 13 March 2026  00:45:46 +0000 (0:00:00.142)       0:00:36.787 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nFriday 13 March 2026  00:45:46 +0000 (0:00:00.131)       0:00:36.918 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.265)       0:00:37.184 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.134)       0:00:37.318 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.124)       0:00:37.443 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.098)       0:00:37.542 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.134)       0:00:37.677 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nFriday 13 March 2026  00:45:47 +0000 (0:00:00.135)       0:00:37.813 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.145)       0:00:37.958 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.136)       0:00:38.094 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.123)       0:00:38.218 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.123)       0:00:38.341 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.128)       0:00:38.470 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.135)       0:00:38.605 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nFriday 13 March 2026  00:45:48 +0000 (0:00:00.287)       0:00:38.893 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nFriday 13 March 2026  00:45:49 +0000 (0:00:00.561)       0:00:39.454 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nFriday 13 March 2026  00:45:50 +0000 (0:00:00.514)       0:00:39.968 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nFriday 13 March 2026  00:45:50 +0000 (0:00:00.529)       0:00:40.498 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nFriday 13 March 2026  00:45:50 +0000 (0:00:00.154)       0:00:40.653 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nFriday 13 March 2026  00:45:50 +0000 (0:00:00.112)       0:00:40.765 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nFriday 13 March 2026  00:45:50 +0000 (0:00:00.118)       0:00:40.884 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.148)       0:00:41.033 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.128)       0:00:41.161 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.129)       0:00:41.291 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.130)       0:00:41.421 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.131)       0:00:41.553 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.248)       0:00:41.802 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nFriday 13 March 2026  00:45:51 +0000 (0:00:00.125)       0:00:41.927 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.126)       0:00:42.054 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.143)       0:00:42.197 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.131)       0:00:42.329 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.137)       0:00:42.467 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.119)       0:00:42.586 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.120)       0:00:42.706 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:45:52 +0000 (0:00:00.110)       0:00:42.816 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.132)       0:00:42.948 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.144)       0:00:43.093 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.139)       0:00:43.232 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.150)       0:00:43.383 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.332)       0:00:43.715 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nFriday 13 March 2026  00:45:53 +0000 (0:00:00.148)       0:00:43.864 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nFriday 13 March 2026  00:45:54 +0000 (0:00:00.135)       0:00:43.999 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nFriday 13 March 2026  00:45:54 +0000 (0:00:00.133)       0:00:44.133 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nFriday 13 March 2026  00:45:54 +0000 (0:00:00.128)       0:00:44.261 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nFriday 13 March 2026  00:45:54 +0000 (0:00:00.503)       0:00:44.764 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nFriday 13 March 2026  00:45:55 +0000 (0:00:00.512)       0:00:45.277 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nFriday 13 March 2026  00:45:55 +0000 (0:00:00.130)       0:00:45.407 ********** \n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'vg_name': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'vg_name': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nFriday 13 March 2026  00:45:55 +0000 (0:00:00.137)       0:00:45.545 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nFriday 13 March 2026  00:45:55 +0000 (0:00:00.131)       0:00:45.676 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nFriday 13 March 2026  00:45:55 +0000 (0:00:00.130)       0:00:45.806 ********** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nFriday 13 March 2026  00:45:56 +0000 (0:00:00.125)       0:00:45.932 ********** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-14d46521-cd63-5c40-b949-080247a26419\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-27536765-3469-5fa1-a30d-2781c44211ac\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  00:45:56 +0000 (0:00:00.379)       0:00:46.311 ********** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nFriday 13 March 2026  00:45:56 +0000 (0:00:00.224)       0:00:46.535 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:56 +0000 (0:00:00.207)       0:00:46.743 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.352)       0:00:47.096 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.145)       0:00:47.242 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.136)       0:00:47.378 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.156)       0:00:47.535 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.165)       0:00:47.700 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:57 +0000 (0:00:00.164)       0:00:47.865 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:58 +0000 (0:00:00.457)       0:00:48.322 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:58 +0000 (0:00:00.180)       0:00:48.503 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:58 +0000 (0:00:00.150)       0:00:48.654 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:59 +0000 (0:00:00.381)       0:00:49.036 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:59 +0000 (0:00:00.379)       0:00:49.415 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:45:59 +0000 (0:00:00.387)       0:00:49.802 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nFriday 13 March 2026  00:46:00 +0000 (0:00:00.393)       0:00:50.196 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:00 +0000 (0:00:00.312)       0:00:50.508 ********** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:00 +0000 (0:00:00.386)       0:00:50.895 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:01 +0000 (0:00:00.181)       0:00:51.077 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:01 +0000 (0:00:00.509)       0:00:51.586 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:01 +0000 (0:00:00.201)       0:00:51.788 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:02 +0000 (0:00:00.174)       0:00:51.963 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:02 +0000 (0:00:00.181)       0:00:52.144 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:02 +0000 (0:00:00.183)       0:00:52.327 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:02 +0000 (0:00:00.178)       0:00:52.506 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:02 +0000 (0:00:00.175)       0:00:52.681 ********** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:03 +0000 (0:00:00.622)       0:00:53.304 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:03 +0000 (0:00:00.204)       0:00:53.508 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:03 +0000 (0:00:00.200)       0:00:53.709 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nFriday 13 March 2026  00:46:03 +0000 (0:00:00.182)       0:00:53.891 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nFriday 13 March 2026  00:46:04 +0000 (0:00:00.185)       0:00:54.077 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nFriday 13 March 2026  00:46:04 +0000 (0:00:00.244)       0:00:54.322 ********** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nFriday 13 March 2026  00:46:04 +0000 (0:00:00.188)       0:00:54.510 ********** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nFriday 13 March 2026  00:46:06 +0000 (0:00:01.806)       0:00:56.317 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nFriday 13 March 2026  00:46:06 +0000 (0:00:00.146)       0:00:56.464 ********** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nFriday 13 March 2026  00:46:07 +0000 (0:00:01.359)       0:00:57.823 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.140)       0:00:57.963 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.122)       0:00:58.086 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.117)       0:00:58.204 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.126)       0:00:58.330 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.120)       0:00:58.451 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.109)       0:00:58.561 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nFriday 13 March 2026  00:46:08 +0000 (0:00:00.142)       0:00:58.703 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.269)       0:00:58.972 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.163)       0:00:59.135 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.143)       0:00:59.279 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.117)       0:00:59.397 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.100)       0:00:59.497 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.111)       0:00:59.609 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.103)       0:00:59.713 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nFriday 13 March 2026  00:46:09 +0000 (0:00:00.131)       0:00:59.844 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nFriday 13 March 2026  00:46:10 +0000 (0:00:00.125)       0:00:59.969 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nFriday 13 March 2026  00:46:10 +0000 (0:00:00.125)       0:01:00.095 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nFriday 13 March 2026  00:46:10 +0000 (0:00:00.544)       0:01:00.640 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nFriday 13 March 2026  00:46:11 +0000 (0:00:00.539)       0:01:01.179 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nFriday 13 March 2026  00:46:11 +0000 (0:00:00.652)       0:01:01.831 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.140)       0:01:01.971 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.098)       0:01:02.070 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.104)       0:01:02.175 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.125)       0:01:02.301 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.112)       0:01:02.413 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.119)       0:01:02.533 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.112)       0:01:02.645 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.109)       0:01:02.755 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nFriday 13 March 2026  00:46:12 +0000 (0:00:00.110)       0:01:02.865 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.123)       0:01:02.988 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.122)       0:01:03.110 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.260)       0:01:03.370 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.123)       0:01:03.494 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.116)       0:01:03.610 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.112)       0:01:03.722 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nFriday 13 March 2026  00:46:13 +0000 (0:00:00.116)       0:01:03.838 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.128)       0:01:03.966 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.124)       0:01:04.091 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.143)       0:01:04.235 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.153)       0:01:04.388 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.138)       0:01:04.527 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.140)       0:01:04.667 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nFriday 13 March 2026  00:46:14 +0000 (0:00:00.132)       0:01:04.800 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nFriday 13 March 2026  00:46:15 +0000 (0:00:00.269)       0:01:05.069 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nFriday 13 March 2026  00:46:15 +0000 (0:00:00.133)       0:01:05.203 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nFriday 13 March 2026  00:46:15 +0000 (0:00:00.127)       0:01:05.331 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nFriday 13 March 2026  00:46:15 +0000 (0:00:00.523)       0:01:05.854 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nFriday 13 March 2026  00:46:16 +0000 (0:00:00.573)       0:01:06.427 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nFriday 13 March 2026  00:46:16 +0000 (0:00:00.153)       0:01:06.581 ********** \n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'vg_name': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'vg_name': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nFriday 13 March 2026  00:46:16 +0000 (0:00:00.148)       0:01:06.729 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nFriday 13 March 2026  00:46:16 +0000 (0:00:00.151)       0:01:06.881 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nFriday 13 March 2026  00:46:17 +0000 (0:00:00.138)       0:01:07.020 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nFriday 13 March 2026  00:46:17 +0000 (0:00:00.136)       0:01:07.156 ********** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-02d8354f-c2be-5171-944a-5a76777090d4\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-a323501f-4860-533e-be9d-aa017f55891f\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:46:17 +0000 (0:00:00.159)       0:01:07.316 ********** \n=============================================================================== \nCreate block VGs -------------------------------------------------------- 5.42s\nCreate block LVs -------------------------------------------------------- 4.16s\nGather DB VGs with total and available size in bytes -------------------- 1.72s\nGather DB+WAL VGs with total and available size in bytes ---------------- 1.68s\nGet list of Ceph PVs with associated VGs -------------------------------- 1.59s\nGather WAL VGs with total and available size in bytes ------------------- 1.57s\nGet list of Ceph LVs with associated VGs -------------------------------- 1.52s\nAdd known partitions to the list of available block devices ------------- 1.34s\nAdd known links to the list of available block devices ------------------ 1.25s\nAdd known links to the list of available block devices ------------------ 1.15s\nAdd known partitions to the list of available block devices ------------- 0.97s\nPrint LVM report data --------------------------------------------------- 0.83s\nGet initial list of available block devices ----------------------------- 0.81s\nAdd known links to the list of available block devices ------------------ 0.77s\nAdd known partitions to the list of available block devices ------------- 0.77s\nGet extra vars for Ceph configuration ----------------------------------- 0.74s\nAdd known partitions to the list of available block devices ------------- 0.66s\nPrint 'Create WAL LVs for ceph_wal_devices' ----------------------------- 0.63s\nAdd known partitions to the list of available block devices ------------- 0.62s\nAdd known links to the list of available block devices ------------------ 0.60s\n\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9b508752-6c86-455f-88c4-f40effbcb7a8 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9b508752-6c86-455f-88c4-f40effbcb7a8 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nFriday 13 March 2026  00:46:33 +0000 (0:00:00.207)       0:00:00.207 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nFriday 13 March 2026  00:46:34 +0000 (0:00:00.970)       0:00:01.178 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  00:46:35 +0000 (0:00:01.089)       0:00:02.267 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nFriday 13 March 2026  00:46:40 +0000 (0:00:05.209)       0:00:07.476 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:46:41 +0000 (0:00:00.448)       0:00:07.925 ********** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.21s\nosism.commons.facts : Copy fact files ----------------------------------- 1.09s\nosism.commons.facts : Create custom facts directory --------------------- 0.97s\nGather facts for all hosts ---------------------------------------------- 0.45s\n\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m\n\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 715d3fd8-4a42-42a9-9503-eff1236b6ac4 (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 715d3fd8-4a42-42a9-9503-eff1236b6ac4 (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nFriday 13 March 2026  00:46:57 +0000 (0:00:00.174)       0:00:00.174 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nFriday 13 March 2026  00:46:57 +0000 (0:00:00.172)       0:00:00.347 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nFriday 13 March 2026  00:46:58 +0000 (0:00:01.067)       0:00:01.414 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nFriday 13 March 2026  00:47:07 +0000 (0:00:09.158)       0:00:10.573 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nFriday 13 March 2026  00:47:08 +0000 (0:00:01.063)       0:00:11.636 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nFriday 13 March 2026  00:47:09 +0000 (0:00:00.933)       0:00:12.570 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Write frr_config_template to temporary file] ********\nFriday 13 March 2026  00:47:10 +0000 (0:00:01.180)       0:00:13.750 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***\nFriday 13 March 2026  00:47:11 +0000 (0:00:00.155)       0:00:13.906 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Remove temporary frr_config_template file] **********\nFriday 13 March 2026  00:47:11 +0000 (0:00:00.142)       0:00:14.049 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nFriday 13 March 2026  00:47:11 +0000 (0:00:00.154)       0:00:14.204 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nFriday 13 March 2026  00:47:11 +0000 (0:00:00.146)       0:00:14.350 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nFriday 13 March 2026  00:47:11 +0000 (0:00:00.143)       0:00:14.494 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nFriday 13 March 2026  00:47:12 +0000 (0:00:01.135)       0:00:15.630 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nFriday 13 March 2026  00:47:15 +0000 (0:00:02.180)       0:00:17.811 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.frr : Restart frr service] *********************\nFriday 13 March 2026  00:47:16 +0000 (0:00:01.055)       0:00:18.866 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:47:17 +0000 (0:00:01.261)       0:00:20.127 ********** \n=============================================================================== \nosism.services.frr : Install frr package -------------------------------- 9.16s\nosism.services.frr : Set sysctl parameters ------------------------------ 2.18s\nosism.services.frr : Restart frr service -------------------------------- 1.26s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 1.18s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.14s\nosism.services.frr : Pin frr package version ---------------------------- 1.07s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.06s\nosism.services.frr : Manage frr service --------------------------------- 1.06s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.93s\nosism.services.frr : Include distribution specific install tasks -------- 0.17s\nosism.services.frr : Write frr_config_template to temporary file -------- 0.16s\nosism.services.frr : Remove temporary frr_config_template file ---------- 0.15s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 0.15s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 0.14s\nosism.services.frr : Render frr.conf from frr_config_template variable --- 0.14s\n\n--> DEPLOY IN A NUTSHELL -- START -- Fri Mar 13 00:47:17 UTC 2026\n\n\u001b[32m2026-03-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m\n\u001b[32m2026-03-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m\n\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m\n\u001b[32m2026-03-13 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\nPLAY [Apply role geerlingguy.dotfiles] *****************************************\n\nTASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****\nFriday 13 March 2026  00:47:42 +0000 (0:00:00.703)       0:00:00.703 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********\nFriday 13 March 2026  00:47:46 +0000 (0:00:04.267)       0:00:04.971 ********** \n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***\nFriday 13 March 2026  00:47:49 +0000 (0:00:02.783)       0:00:07.755 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.730324', 'end': '2026-03-13 00:47:47.735404', 'delta': '0:00:00.005080', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.738730', 'end': '2026-03-13 00:47:47.742517', 'delta': '0:00:00.003787', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.732844', 'end': '2026-03-13 00:47:47.737422', 'delta': '0:00:00.004578', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.813782', 'end': '2026-03-13 00:47:47.818180', 'delta': '0:00:00.004398', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.778395', 'end': '2026-03-13 00:47:48.788111', 'delta': '0:00:00.009716', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.095443', 'end': '2026-03-13 00:47:48.099448', 'delta': '0:00:00.004005', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.744434', 'end': '2026-03-13 00:47:48.749859', 'delta': '0:00:00.005425', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****\nFriday 13 March 2026  00:47:52 +0000 (0:00:03.177)       0:00:10.932 ********** \n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************\nFriday 13 March 2026  00:47:54 +0000 (0:00:01.708)       0:00:12.641 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:47:56 +0000 (0:00:01.905)       0:00:14.547 ********** \n=============================================================================== \ngeerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 4.27s\ngeerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 3.18s\ngeerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.78s\ngeerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 1.91s\ngeerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.71s\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role homer] ********************************************************\n\nTASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***\nFriday 13 March 2026  00:47:41 +0000 (0:00:00.372)       0:00:00.372 ********** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.services.homer : Create traefik external network] ******************\nFriday 13 March 2026  00:47:41 +0000 (0:00:00.245)       0:00:00.617 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Create required directories] **********************\nFriday 13 March 2026  00:47:43 +0000 (0:00:02.172)       0:00:02.794 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m\n\nTASK [osism.services.homer : Copy config.yml configuration file] ***************\nFriday 13 March 2026  00:47:45 +0000 (0:00:01.778)       0:00:04.573 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Copy docker-compose.yml file] *********************\nFriday 13 March 2026  00:47:49 +0000 (0:00:04.452)       0:00:09.025 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Manage homer service] *****************************\nFriday 13 March 2026  00:47:52 +0000 (0:00:02.766)       0:00:11.792 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.homer : Restart homer service] *****************\nFriday 13 March 2026  00:48:19 +0000 (0:00:26.460)       0:00:38.252 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:48:21 +0000 (0:00:02.572)       0:00:40.825 ********** \n=============================================================================== \nosism.services.homer : Manage homer service ---------------------------- 26.46s\nosism.services.homer : Copy config.yml configuration file --------------- 4.45s\nosism.services.homer : Copy docker-compose.yml file --------------------- 2.77s\nosism.services.homer : Restart homer service ---------------------------- 2.57s\nosism.services.homer : Create traefik external network ------------------ 2.18s\nosism.services.homer : Create required directories ---------------------- 1.78s\nosism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.25s\n\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nFriday 13 March 2026  00:47:44 +0000 (0:00:01.482)       0:00:01.482 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nFriday 13 March 2026  00:47:45 +0000 (0:00:00.791)       0:00:02.274 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nFriday 13 March 2026  00:47:46 +0000 (0:00:01.756)       0:00:04.030 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nFriday 13 March 2026  00:47:52 +0000 (0:00:05.800)       0:00:09.830 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nFriday 13 March 2026  00:48:26 +0000 (0:00:33.629)       0:00:43.459 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nFriday 13 March 2026  00:48:27 +0000 (0:00:00.982)       0:00:44.442 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nFriday 13 March 2026  00:48:27 +0000 (0:00:00.532)       0:00:44.975 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nFriday 13 March 2026  00:48:29 +0000 (0:00:01.657)       0:00:46.632 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nFriday 13 March 2026  00:48:30 +0000 (0:00:00.615)       0:00:47.247 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nFriday 13 March 2026  00:48:30 +0000 (0:00:00.671)       0:00:47.919 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:48:31 +0000 (0:00:00.432)       0:00:48.351 ********** \n=============================================================================== \nosism.services.openstackclient : Manage openstackclient service -------- 33.63s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 5.80s\nosism.services.openstackclient : Create required directories ------------ 1.76s\nosism.services.openstackclient : Restart openstackclient service -------- 1.66s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 0.98s\nosism.services.openstackclient : Include tasks -------------------------- 0.79s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.67s\nosism.services.openstackclient : Ensure that all containers are up ------ 0.62s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 0.53s\nosism.services.openstackclient : Copy bash completion script ------------ 0.43s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:47:41 +0000 (0:00:00.532)       0:00:00.532 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m\n\nPLAY [Apply role netdata] ******************************************************\n\nTASK [osism.services.netdata : Include distribution specific install tasks] ****\nFriday 13 March 2026  00:47:44 +0000 (0:00:02.999)       0:00:03.532 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Remove old architecture-dependent repository] ***\nFriday 13 March 2026  00:47:46 +0000 (0:00:01.242)       0:00:04.774 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Install apt-transport-https package] ************\nFriday 13 March 2026  00:47:47 +0000 (0:00:01.535)       0:00:06.309 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Add repository gpg key] *************************\nFriday 13 March 2026  00:47:50 +0000 (0:00:02.886)       0:00:09.196 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add repository] *********************************\nFriday 13 March 2026  00:47:53 +0000 (0:00:02.593)       0:00:11.790 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.netdata : Install package netdata] ************************\nFriday 13 March 2026  00:48:03 +0000 (0:00:10.097)       0:00:21.887 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.netdata : Include config tasks] ***************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:37.139)       0:00:59.027 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Copy configuration files] ***********************\nFriday 13 March 2026  00:48:41 +0000 (0:00:01.399)       0:01:00.426 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m\n\nTASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***\nFriday 13 March 2026  00:48:46 +0000 (0:00:05.231)       0:01:05.657 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Opt out from anonymous statistics] **************\nFriday 13 March 2026  00:48:48 +0000 (0:00:01.830)       0:01:07.488 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add netdata user to docker group] ***************\nFriday 13 March 2026  00:48:51 +0000 (0:00:02.267)       0:01:09.756 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Manage service netdata] *************************\nFriday 13 March 2026  00:48:54 +0000 (0:00:03.529)       0:01:13.285 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Include host type specific tasks] ***************\nFriday 13 March 2026  00:48:57 +0000 (0:00:02.470)       0:01:15.756 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********\nFriday 13 March 2026  00:48:58 +0000 (0:00:01.354)       0:01:17.111 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.netdata : Restart service netdata] *************\nFriday 13 March 2026  00:49:00 +0000 (0:00:01.871)       0:01:18.982 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:49:11 +0000 (0:00:11.373)       0:01:30.356 ********** \n=============================================================================== \nosism.services.netdata : Install package netdata ----------------------- 37.14s\nosism.services.netdata : Restart service netdata ----------------------- 11.37s\nosism.services.netdata : Add repository -------------------------------- 10.10s\nosism.services.netdata : Copy configuration files ----------------------- 5.23s\nosism.services.netdata : Add netdata user to docker group --------------- 3.53s\nGroup hosts based on enabled services ----------------------------------- 3.00s\nosism.services.netdata : Install apt-transport-https package ------------ 2.89s\nosism.services.netdata : Add repository gpg key ------------------------- 2.59s\nosism.services.netdata : Manage service netdata ------------------------- 2.47s\nosism.services.netdata : Opt out from anonymous statistics -------------- 2.27s\nosism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.87s\nosism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.83s\nosism.services.netdata : Remove old architecture-dependent repository --- 1.54s\nosism.services.netdata : Include config tasks --------------------------- 1.40s\nosism.services.netdata : Include host type specific tasks --------------- 1.36s\nosism.services.netdata : Include distribution specific install tasks ---- 1.24s\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role phpmyadmin] ***************************************************\n\nTASK [osism.services.phpmyadmin : Create traefik external network] *************\nFriday 13 March 2026  00:48:03 +0000 (0:00:00.253)       0:00:00.253 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Create required directories] *****************\nFriday 13 March 2026  00:48:05 +0000 (0:00:02.240)       0:00:02.494 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m\n\nTASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************\nFriday 13 March 2026  00:48:06 +0000 (0:00:01.003)       0:00:03.500 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************\nFriday 13 March 2026  00:48:08 +0000 (0:00:02.043)       0:00:05.543 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******\nFriday 13 March 2026  00:49:22 +0000 (0:01:14.327)       0:01:19.870 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:49:26 +0000 (0:00:03.764)       0:01:23.635 ********** \n=============================================================================== \nosism.services.phpmyadmin : Manage phpmyadmin service ------------------ 74.33s\nosism.services.phpmyadmin : Restart phpmyadmin service ------------------ 3.76s\nosism.services.phpmyadmin : Create traefik external network ------------- 2.24s\nosism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 2.04s\nosism.services.phpmyadmin : Create required directories ----------------- 1.01s\n\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nFriday 13 March 2026  00:47:34 +0000 (0:00:00.232)       0:00:00.232 ********** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nFriday 13 March 2026  00:47:35 +0000 (0:00:01.245)       0:00:01.478 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nFriday 13 March 2026  00:47:40 +0000 (0:00:04.305)       0:00:05.783 ********** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nFriday 13 March 2026  00:47:41 +0000 (0:00:01.483)       0:00:07.267 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:47:47 +0000 (0:00:05.518)       0:00:12.785 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nFriday 13 March 2026  00:47:50 +0000 (0:00:02.791)       0:00:15.577 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensure /var/log/journal exists on EL10 systems] *****************\nFriday 13 March 2026  00:47:54 +0000 (0:00:04.512)       0:00:20.089 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nFriday 13 March 2026  00:47:56 +0000 (0:00:01.742)       0:00:21.832 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nFriday 13 March 2026  00:47:58 +0000 (0:00:01.754)       0:00:23.586 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over kolla.target] **************************************\nFriday 13 March 2026  00:47:59 +0000 (0:00:01.285)       0:00:24.872 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nFriday 13 March 2026  00:48:02 +0000 (0:00:03.573)       0:00:28.446 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nFriday 13 March 2026  00:48:10 +0000 (0:00:07.182)       0:00:35.628 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nFriday 13 March 2026  00:48:11 +0000 (0:00:01.050)       0:00:36.679 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nFriday 13 March 2026  00:48:11 +0000 (0:00:00.744)       0:00:37.423 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nFriday 13 March 2026  00:48:12 +0000 (0:00:00.775)       0:00:38.199 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nFriday 13 March 2026  00:48:13 +0000 (0:00:00.818)       0:00:39.017 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nFriday 13 March 2026  00:48:18 +0000 (0:00:05.011)       0:00:44.029 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nFriday 13 March 2026  00:48:21 +0000 (0:00:03.393)       0:00:47.423 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nFriday 13 March 2026  00:48:24 +0000 (0:00:02.172)       0:00:49.596 ********** \n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nFriday 13 March 2026  00:48:25 +0000 (0:00:01.748)       0:00:51.344 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nFriday 13 March 2026  00:48:27 +0000 (0:00:01.896)       0:00:53.241 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [service-check-containers : common | Check containers] ********************\nFriday 13 March 2026  00:48:30 +0000 (0:00:02.342)       0:00:55.584 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : common | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:48:33 +0000 (0:00:03.435)       0:00:59.019 ********** \n\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:48:34 +0000 (0:00:01.085)       0:01:00.105 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Creating log volume] ********************************************\nFriday 13 March 2026  00:48:37 +0000 (0:00:02.644)       0:01:02.749 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Link kolla_logs volume to /var/log/kolla] ***********************\nFriday 13 March 2026  00:48:38 +0000 (0:00:01.756)       0:01:04.505 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:01.239)       0:01:05.745 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.089)       0:01:05.834 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.062)       0:01:05.897 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.199)       0:01:06.097 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.059)       0:01:06.156 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.075)       0:01:06.232 ********** \n\nTASK [common : Flush handlers] *************************************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.085)       0:01:06.317 ********** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nFriday 13 March 2026  00:48:40 +0000 (0:00:00.110)       0:01:06.428 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nFriday 13 March 2026  00:49:17 +0000 (0:00:36.775)       0:01:43.204 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nFriday 13 March 2026  00:49:50 +0000 (0:00:33.262)       0:02:16.467 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nFriday 13 March 2026  00:49:52 +0000 (0:00:02.086)       0:02:18.554 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:50:02 +0000 (0:00:09.264)       0:02:27.818 ********** \n=============================================================================== \ncommon : Restart fluentd container ------------------------------------- 36.78s\ncommon : Restart kolla-toolbox container ------------------------------- 33.26s\ncommon : Restart cron container ----------------------------------------- 9.26s\ncommon : Copying over config.json files for services -------------------- 7.18s\nservice-cert-copy : common | Copying over extra CA certificates --------- 5.52s\ncommon : Copying over fluentd.conf -------------------------------------- 5.01s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 4.51s\ncommon : Ensuring config directories exist ------------------------------ 4.31s\ncommon : Copying over kolla.target -------------------------------------- 3.57s\nservice-check-containers : common | Check containers -------------------- 3.44s\ncommon : Copying over cron logrotate config file ------------------------ 3.39s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 2.79s\nservice-check-containers : Include tasks -------------------------------- 2.64s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.34s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.17s\ncommon : Initializing toolbox container using normal user --------------- 2.09s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 1.90s\ncommon : Creating log volume -------------------------------------------- 1.76s\ncommon : Copying over /run subdirectories conf -------------------------- 1.75s\ncommon : Ensuring config directories have correct owner and permission --- 1.75s\n\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.407)       0:00:00.407 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.322)       0:00:00.730 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.466)       0:00:01.196 ********** \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.631)       0:00:01.828 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nFriday 13 March 2026  00:50:11 +0000 (0:00:00.853)       0:00:02.681 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\nTASK [service-check-containers : memcached | Check containers] *****************\nFriday 13 March 2026  00:50:13 +0000 (0:00:02.193)       0:00:04.875 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [service-check-containers : memcached | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:50:15 +0000 (0:00:01.419)       0:00:06.295 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:50:15 +0000 (0:00:00.455)       0:00:06.750 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nFriday 13 March 2026  00:50:17 +0000 (0:00:01.801)       0:00:08.551 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:50:24 +0000 (0:00:07.280)       0:00:15.832 ********** \n=============================================================================== \nmemcached : Restart memcached container --------------------------------- 7.28s\nmemcached : Copying over config.json files for services ----------------- 2.19s\nservice-check-containers : Include tasks -------------------------------- 1.80s\nservice-check-containers : memcached | Check containers ----------------- 1.42s\nmemcached : Ensuring config directories exist --------------------------- 0.85s\nmemcached : include_tasks ----------------------------------------------- 0.63s\nGroup hosts based on enabled services ----------------------------------- 0.47s\nservice-check-containers : memcached | Notify handlers to restart containers --- 0.46s\nGroup hosts based on Kolla action --------------------------------------- 0.32s\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:50:08 +0000 (0:00:00.290)       0:00:00.290 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.381)       0:00:00.671 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.564)       0:00:01.236 ********** \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.717)       0:00:01.953 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nFriday 13 March 2026  00:50:11 +0000 (0:00:01.412)       0:00:03.365 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nFriday 13 March 2026  00:50:14 +0000 (0:00:02.709)       0:00:06.075 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : redis | Check containers] *********************\nFriday 13 March 2026  00:50:17 +0000 (0:00:02.917)       0:00:08.993 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : redis | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:50:19 +0000 (0:00:01.852)       0:00:10.846 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:50:19 +0000 (0:00:00.469)       0:00:11.315 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nFriday 13 March 2026  00:50:21 +0000 (0:00:01.112)       0:00:12.427 ********** \n\nTASK [redis : Flush handlers] **************************************************\nFriday 13 March 2026  00:50:21 +0000 (0:00:00.112)       0:00:12.540 ********** \n\nTASK [redis : Flush handlers] **************************************************\nFriday 13 March 2026  00:50:21 +0000 (0:00:00.175)       0:00:12.716 ********** \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nFriday 13 March 2026  00:50:21 +0000 (0:00:00.124)       0:00:12.840 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nFriday 13 March 2026  00:50:24 +0000 (0:00:02.758)       0:00:15.599 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:50:28 +0000 (0:00:04.700)       0:00:20.299 ********** \n=============================================================================== \nredis : Restart redis-sentinel container -------------------------------- 4.70s\nredis : Copying over redis config files --------------------------------- 2.92s\nredis : Restart redis container ----------------------------------------- 2.76s\nredis : Copying over default config.json files -------------------------- 2.71s\nservice-check-containers : redis | Check containers --------------------- 1.85s\nredis : Ensuring config directories exist ------------------------------- 1.41s\nservice-check-containers : Include tasks -------------------------------- 1.11s\nredis : include_tasks --------------------------------------------------- 0.72s\nGroup hosts based on enabled services ----------------------------------- 0.56s\nservice-check-containers : redis | Notify handlers to restart containers --- 0.47s\nredis : Flush handlers -------------------------------------------------- 0.41s\nGroup hosts based on Kolla action --------------------------------------- 0.38s\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:50:08 +0000 (0:00:00.226)       0:00:00.226 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.995)       0:00:01.221 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.670)       0:00:01.892 ********** \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nFriday 13 March 2026  00:50:11 +0000 (0:00:01.360)       0:00:03.253 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nFriday 13 March 2026  00:50:13 +0000 (0:00:01.460)       0:00:04.714 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nFriday 13 March 2026  00:50:15 +0000 (0:00:01.901)       0:00:06.615 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nFriday 13 March 2026  00:50:16 +0000 (0:00:01.774)       0:00:08.390 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nFriday 13 March 2026  00:50:18 +0000 (0:00:01.588)       0:00:09.978 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nFriday 13 March 2026  00:50:20 +0000 (0:00:01.693)       0:00:11.672 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nFriday 13 March 2026  00:50:22 +0000 (0:00:02.569)       0:00:14.242 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-check-containers : openvswitch | Check containers] ***************\nFriday 13 March 2026  00:50:23 +0000 (0:00:00.835)       0:00:15.078 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : openvswitch | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:50:26 +0000 (0:00:03.176)       0:00:18.255 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:50:27 +0000 (0:00:01.014)       0:00:19.269 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:30 +0000 (0:00:02.418)       0:00:21.687 ********** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:30 +0000 (0:00:00.288)       0:00:21.976 ********** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:30 +0000 (0:00:00.195)       0:00:22.171 ********** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:30 +0000 (0:00:00.156)       0:00:22.328 ********** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.486)       0:00:22.814 ********** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.198)       0:00:23.012 ********** \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.169)       0:00:23.182 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nFriday 13 March 2026  00:50:37 +0000 (0:00:05.232)       0:00:28.414 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nFriday 13 March 2026  00:50:38 +0000 (0:00:01.571)       0:00:29.986 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nFriday 13 March 2026  00:50:47 +0000 (0:00:08.721)       0:00:38.707 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nFriday 13 March 2026  00:50:55 +0000 (0:00:07.797)       0:00:46.504 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nFriday 13 March 2026  00:50:57 +0000 (0:00:02.081)       0:00:48.586 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nFriday 13 March 2026  00:51:00 +0000 (0:00:03.552)       0:00:52.138 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:51:09 +0000 (0:00:08.398)       0:01:00.537 ********** \n=============================================================================== \nopenvswitch : Restart openvswitch-vswitchd container ------------------- 17.12s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 7.80s\nopenvswitch : Restart openvswitch-db-server container ------------------- 5.23s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 3.55s\nservice-check-containers : openvswitch | Check containers --------------- 3.18s\nopenvswitch : Copying over config.json files for services --------------- 2.57s\nservice-check-containers : Include tasks -------------------------------- 2.42s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 2.08s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.90s\nmodule-load : Drop module persistence ----------------------------------- 1.77s\nopenvswitch : Ensuring config directories exist ------------------------- 1.69s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 1.59s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 1.57s\nopenvswitch : Flush Handlers -------------------------------------------- 1.49s\nmodule-load : Load modules ---------------------------------------------- 1.46s\nopenvswitch : include_tasks --------------------------------------------- 1.36s\nservice-check-containers : openvswitch | Notify handlers to restart containers --- 1.01s\nGroup hosts based on Kolla action --------------------------------------- 1.00s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.84s\nGroup hosts based on enabled services ----------------------------------- 0.67s\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state SUCCESS\u001b[0m\n\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nFriday 13 March 2026  00:47:34 +0000 (0:00:00.216)       0:00:00.216 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nFriday 13 March 2026  00:47:35 +0000 (0:00:00.696)       0:00:00.912 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nFriday 13 March 2026  00:47:36 +0000 (0:00:00.567)       0:00:01.480 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nFriday 13 March 2026  00:47:36 +0000 (0:00:00.636)       0:00:02.116 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nFriday 13 March 2026  00:47:39 +0000 (0:00:02.605)       0:00:04.722 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nFriday 13 March 2026  00:47:41 +0000 (0:00:02.356)       0:00:07.078 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nFriday 13 March 2026  00:47:42 +0000 (0:00:01.101)       0:00:08.179 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nFriday 13 March 2026  00:47:43 +0000 (0:00:00.911)       0:00:09.091 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nFriday 13 March 2026  00:47:44 +0000 (0:00:00.773)       0:00:09.865 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nFriday 13 March 2026  00:47:45 +0000 (0:00:00.928)       0:00:10.794 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nFriday 13 March 2026  00:47:47 +0000 (0:00:01.644)       0:00:12.438 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nFriday 13 March 2026  00:47:48 +0000 (0:00:01.069)       0:00:13.508 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nFriday 13 March 2026  00:47:53 +0000 (0:00:05.213)       0:00:18.721 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nFriday 13 March 2026  00:47:55 +0000 (0:00:02.053)       0:00:20.775 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nFriday 13 March 2026  00:47:57 +0000 (0:00:02.361)       0:00:23.137 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nFriday 13 March 2026  00:47:59 +0000 (0:00:01.139)       0:00:24.277 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nFriday 13 March 2026  00:48:00 +0000 (0:00:01.641)       0:00:25.919 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nFriday 13 March 2026  00:48:02 +0000 (0:00:01.477)       0:00:27.396 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nFriday 13 March 2026  00:48:04 +0000 (0:00:02.600)       0:00:29.996 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nFriday 13 March 2026  00:48:06 +0000 (0:00:01.977)       0:00:31.974 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nFriday 13 March 2026  00:48:08 +0000 (0:00:01.464)       0:00:33.438 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nFriday 13 March 2026  00:48:09 +0000 (0:00:01.119)       0:00:34.558 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nFriday 13 March 2026  00:48:10 +0000 (0:00:00.792)       0:00:35.351 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nFriday 13 March 2026  00:48:10 +0000 (0:00:00.312)       0:00:35.664 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nFriday 13 March 2026  00:48:11 +0000 (0:00:01.312)       0:00:36.976 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nFriday 13 March 2026  00:48:13 +0000 (0:00:01.305)       0:00:38.282 ********** \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nFriday 13 March 2026  00:48:13 +0000 (0:00:00.455)       0:00:38.738 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nFriday 13 March 2026  00:48:16 +0000 (0:00:02.938)       0:00:41.676 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nFriday 13 March 2026  00:48:17 +0000 (0:00:00.817)       0:00:42.494 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nFriday 13 March 2026  00:48:18 +0000 (0:00:00.886)       0:00:43.381 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nFriday 13 March 2026  00:48:19 +0000 (0:00:01.286)       0:00:44.668 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nFriday 13 March 2026  00:48:20 +0000 (0:00:00.622)       0:00:45.291 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nFriday 13 March 2026  00:48:20 +0000 (0:00:00.375)       0:00:45.666 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nFriday 13 March 2026  00:48:21 +0000 (0:00:01.524)       0:00:47.191 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nFriday 13 March 2026  00:48:24 +0000 (0:00:02.395)       0:00:49.587 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nFriday 13 March 2026  00:48:24 +0000 (0:00:00.530)       0:00:50.118 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nFriday 13 March 2026  00:49:08 +0000 (0:00:43.418)       0:01:33.537 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nFriday 13 March 2026  00:49:08 +0000 (0:00:00.341)       0:01:33.879 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nFriday 13 March 2026  00:49:09 +0000 (0:00:00.912)       0:01:34.791 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nFriday 13 March 2026  00:49:10 +0000 (0:00:01.367)       0:01:36.158 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nFriday 13 March 2026  00:49:37 +0000 (0:00:26.118)       0:02:02.277 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nFriday 13 March 2026  00:49:37 +0000 (0:00:00.706)       0:02:02.983 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nFriday 13 March 2026  00:49:38 +0000 (0:00:00.651)       0:02:03.634 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nFriday 13 March 2026  00:49:38 +0000 (0:00:00.551)       0:02:04.186 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nFriday 13 March 2026  00:49:39 +0000 (0:00:00.893)       0:02:05.079 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nFriday 13 March 2026  00:49:40 +0000 (0:00:00.286)       0:02:05.366 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nFriday 13 March 2026  00:49:40 +0000 (0:00:00.638)       0:02:06.005 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nFriday 13 March 2026  00:49:41 +0000 (0:00:00.686)       0:02:06.691 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nFriday 13 March 2026  00:49:42 +0000 (0:00:01.194)       0:02:07.886 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nFriday 13 March 2026  00:49:43 +0000 (0:00:01.049)       0:02:08.935 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nFriday 13 March 2026  00:49:43 +0000 (0:00:00.276)       0:02:09.211 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nFriday 13 March 2026  00:49:44 +0000 (0:00:00.269)       0:02:09.480 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nFriday 13 March 2026  00:49:45 +0000 (0:00:00.858)       0:02:10.338 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nFriday 13 March 2026  00:49:45 +0000 (0:00:00.616)       0:02:10.955 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nFriday 13 March 2026  00:49:48 +0000 (0:00:02.758)       0:02:13.714 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nFriday 13 March 2026  00:49:48 +0000 (0:00:00.523)       0:02:14.237 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nFriday 13 March 2026  00:49:49 +0000 (0:00:00.590)       0:02:14.828 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nFriday 13 March 2026  00:49:49 +0000 (0:00:00.283)       0:02:15.112 ********** \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nFriday 13 March 2026  00:49:50 +0000 (0:00:00.542)       0:02:15.655 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nFriday 13 March 2026  00:49:50 +0000 (0:00:00.258)       0:02:15.913 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nFriday 13 March 2026  00:49:50 +0000 (0:00:00.259)       0:02:16.173 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nFriday 13 March 2026  00:49:51 +0000 (0:00:00.324)       0:02:16.497 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nFriday 13 March 2026  00:49:52 +0000 (0:00:01.015)       0:02:17.512 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nFriday 13 March 2026  00:49:53 +0000 (0:00:01.016)       0:02:18.528 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nFriday 13 March 2026  00:49:54 +0000 (0:00:01.320)       0:02:19.848 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nFriday 13 March 2026  00:50:04 +0000 (0:00:09.792)       0:02:29.640 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nFriday 13 March 2026  00:50:05 +0000 (0:00:00.696)       0:02:30.337 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nFriday 13 March 2026  00:50:05 +0000 (0:00:00.500)       0:02:30.838 ********** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nFriday 13 March 2026  00:50:06 +0000 (0:00:00.585)       0:02:31.423 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nFriday 13 March 2026  00:50:07 +0000 (0:00:00.840)       0:02:32.264 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nFriday 13 March 2026  00:50:07 +0000 (0:00:00.984)       0:02:33.248 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nFriday 13 March 2026  00:50:09 +0000 (0:00:01.256)       0:02:34.505 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.657)       0:02:35.162 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.472)       0:02:35.635 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.346)       0:02:35.982 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.127)       0:02:36.109 ********** \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nFriday 13 March 2026  00:50:11 +0000 (0:00:00.207)       0:02:36.316 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nFriday 13 March 2026  00:50:11 +0000 (0:00:00.674)       0:02:36.990 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nFriday 13 March 2026  00:50:13 +0000 (0:00:01.407)       0:02:38.398 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nFriday 13 March 2026  00:50:13 +0000 (0:00:00.802)       0:02:39.201 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nFriday 13 March 2026  00:50:14 +0000 (0:00:00.514)       0:02:39.716 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nFriday 13 March 2026  00:50:20 +0000 (0:00:06.379)       0:02:46.095 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nFriday 13 March 2026  00:50:34 +0000 (0:00:13.686)       0:02:59.781 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.484)       0:03:00.266 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.283)       0:03:00.549 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.325)       0:03:00.874 ********** \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nFriday 13 March 2026  00:50:36 +0000 (0:00:00.574)       0:03:01.449 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nFriday 13 March 2026  00:50:36 +0000 (0:00:00.781)       0:03:02.231 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nFriday 13 March 2026  00:50:37 +0000 (0:00:00.798)       0:03:03.029 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nFriday 13 March 2026  00:50:37 +0000 (0:00:00.103)       0:03:03.132 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nFriday 13 March 2026  00:50:38 +0000 (0:00:00.967)       0:03:04.100 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nFriday 13 March 2026  00:50:38 +0000 (0:00:00.115)       0:03:04.215 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nFriday 13 March 2026  00:50:39 +0000 (0:00:00.112)       0:03:04.327 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nFriday 13 March 2026  00:50:39 +0000 (0:00:00.114)       0:03:04.442 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nFriday 13 March 2026  00:50:39 +0000 (0:00:00.120)       0:03:04.563 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nFriday 13 March 2026  00:50:43 +0000 (0:00:04.216)       0:03:08.779 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nFriday 13 March 2026  00:51:30 +0000 (0:00:46.684)       0:03:55.464 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nFriday 13 March 2026  00:51:31 +0000 (0:00:01.129)       0:03:56.594 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nFriday 13 March 2026  00:51:32 +0000 (0:00:01.597)       0:03:58.191 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nFriday 13 March 2026  00:51:33 +0000 (0:00:00.904)       0:03:59.096 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nFriday 13 March 2026  00:51:33 +0000 (0:00:00.112)       0:03:59.208 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nFriday 13 March 2026  00:51:35 +0000 (0:00:01.844)       0:04:01.052 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nFriday 13 March 2026  00:51:36 +0000 (0:00:00.278)       0:04:01.331 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nFriday 13 March 2026  00:51:37 +0000 (0:00:01.036)       0:04:02.368 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nFriday 13 March 2026  00:51:37 +0000 (0:00:00.121)       0:04:02.489 ********** \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nFriday 13 March 2026  00:51:37 +0000 (0:00:00.207)       0:04:02.696 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nFriday 13 March 2026  00:51:42 +0000 (0:00:05.237)       0:04:07.933 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nFriday 13 March 2026  00:51:43 +0000 (0:00:00.774)       0:04:08.708 ********** \n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nFriday 13 March 2026  00:51:53 +0000 (0:00:09.607)       0:04:18.315 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nFriday 13 March 2026  00:51:53 +0000 (0:00:00.703)       0:04:19.019 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:51:54 +0000 (0:00:00.459)       0:04:19.478 ********** \n=============================================================================== \nk3s_server_post : Wait for Cilium resources ---------------------------- 46.68s\nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.42s\nk3s_server : Enable and check K3s service ------------------------------ 26.12s\nkubectl : Install required packages ------------------------------------ 13.69s\nk3s_agent : Manage k3s service ------------------------------------------ 9.79s\nManage labels ----------------------------------------------------------- 9.61s\nkubectl : Add repository Debian ----------------------------------------- 6.38s\nk9s : Install k9s packages ---------------------------------------------- 5.24s\nk3s_download : Download k3s binary x64 ---------------------------------- 5.21s\nk3s_server_post : Install Cilium ---------------------------------------- 4.22s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.94s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.76s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.61s\nk3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 2.60s\nk3s_server : Detect Kubernetes version for label compatibility ---------- 2.40s\nk3s_download : Download k3s binary armhf -------------------------------- 2.36s\nk3s_prereq : Enable IPv6 forwarding ------------------------------------- 2.36s\nk3s_download : Download k3s binary arm64 -------------------------------- 2.05s\nk3s_server : Validating arguments against arg spec 'main' - Setup k3s servers --- 1.98s\nk3s_server_post : Test for BGP config resources ------------------------- 1.84s\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5dd18988-6afa-48f6-b9a5-301792612dd3 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5dd18988-6afa-48f6-b9a5-301792612dd3 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy kubeconfig to the configuration repository] *************************\n\nTASK [Get kubeconfig file] *****************************************************\nFriday 13 March 2026  00:51:58 +0000 (0:00:00.141)       0:00:00.141 ********** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nFriday 13 March 2026  00:51:59 +0000 (0:00:00.634)       0:00:00.775 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig file] ****************************\nFriday 13 March 2026  00:52:00 +0000 (0:00:01.017)       0:00:01.793 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:52:00 +0000 (0:00:00.437)       0:00:02.230 ********** \n=============================================================================== \nWrite kubeconfig file --------------------------------------------------- 1.02s\nGet kubeconfig file ----------------------------------------------------- 0.63s\nChange server address in the kubeconfig file ---------------------------- 0.44s\n\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nFriday 13 March 2026  00:51:58 +0000 (0:00:00.119)       0:00:00.119 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nFriday 13 March 2026  00:51:58 +0000 (0:00:00.430)       0:00:00.550 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nFriday 13 March 2026  00:51:59 +0000 (0:00:00.506)       0:00:01.056 ********** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nFriday 13 March 2026  00:52:00 +0000 (0:00:00.655)       0:00:01.712 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nFriday 13 March 2026  00:52:01 +0000 (0:00:01.218)       0:00:02.931 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nFriday 13 March 2026  00:52:01 +0000 (0:00:00.509)       0:00:03.440 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nFriday 13 March 2026  00:52:03 +0000 (0:00:01.937)       0:00:05.378 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nFriday 13 March 2026  00:52:04 +0000 (0:00:01.081)       0:00:06.459 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nFriday 13 March 2026  00:52:05 +0000 (0:00:00.414)       0:00:06.874 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:52:05 +0000 (0:00:00.328)       0:00:07.202 ********** \n=============================================================================== \nMake kubeconfig available for use inside the manager service ------------ 1.94s\nWrite kubeconfig file --------------------------------------------------- 1.22s\nChange server address in the kubeconfig inside the manager service ------ 1.08s\nGet kubeconfig file ----------------------------------------------------- 0.66s\nChange server address in the kubeconfig --------------------------------- 0.51s\nCreate .kube directory -------------------------------------------------- 0.51s\nGet home directory of operator user ------------------------------------- 0.43s\nSet KUBECONFIG environment variable ------------------------------------- 0.41s\nEnable kubectl command line completion ---------------------------------- 0.33s\n\n\nPLAY [Set kolla_action_rabbitmq] ***********************************************\n\nTASK [Inform the user about the following task] ********************************\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.148)       0:00:00.148 ********** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check RabbitMQ service] **************************************************\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.064)       0:00:00.212 ********** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******\nFriday 13 March 2026  00:50:35 +0000 (0:00:04.261)       0:00:04.474 ********** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.070)       0:00:04.545 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.243)       0:00:04.788 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:50:36 +0000 (0:00:00.560)       0:00:05.349 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nFriday 13 March 2026  00:50:37 +0000 (0:00:00.931)       0:00:06.281 ********** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nFriday 13 March 2026  00:50:37 +0000 (0:00:00.455)       0:00:06.736 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nFriday 13 March 2026  00:50:39 +0000 (0:00:01.529)       0:00:08.266 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nFriday 13 March 2026  00:50:40 +0000 (0:00:00.957)       0:00:09.224 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nFriday 13 March 2026  00:50:41 +0000 (0:00:01.002)       0:00:10.226 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nFriday 13 March 2026  00:50:41 +0000 (0:00:00.236)       0:00:10.463 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nFriday 13 March 2026  00:50:42 +0000 (0:00:00.646)       0:00:11.109 ********** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nFriday 13 March 2026  00:50:42 +0000 (0:00:00.496)       0:00:11.605 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nFriday 13 March 2026  00:50:43 +0000 (0:00:00.751)       0:00:12.357 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nFriday 13 March 2026  00:50:43 +0000 (0:00:00.277)       0:00:12.635 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nFriday 13 March 2026  00:50:44 +0000 (0:00:00.560)       0:00:13.195 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nFriday 13 March 2026  00:50:45 +0000 (0:00:00.995)       0:00:14.191 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nFriday 13 March 2026  00:50:46 +0000 (0:00:01.483)       0:00:15.674 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nFriday 13 March 2026  00:50:48 +0000 (0:00:01.506)       0:00:17.180 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nFriday 13 March 2026  00:50:50 +0000 (0:00:02.365)       0:00:19.546 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nFriday 13 March 2026  00:50:51 +0000 (0:00:01.372)       0:00:20.919 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nFriday 13 March 2026  00:50:53 +0000 (0:00:01.788)       0:00:22.708 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nFriday 13 March 2026  00:50:54 +0000 (0:00:01.166)       0:00:23.874 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nFriday 13 March 2026  00:50:56 +0000 (0:00:01.625)       0:00:25.500 ********** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over extra CA certificates] *******\nFriday 13 March 2026  00:50:57 +0000 (0:00:00.751)       0:00:26.251 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:50:58 +0000 (0:00:01.453)       0:00:27.705 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : rabbitmq | Copying over backend internal TLS key] ****\nFriday 13 March 2026  00:50:59 +0000 (0:00:00.456)       0:00:28.161 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : rabbitmq | Check containers] ******************\nFriday 13 March 2026  00:50:59 +0000 (0:00:00.808)       0:00:28.969 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [service-check-containers : rabbitmq | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:51:01 +0000 (0:00:01.165)       0:00:30.135 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:51:01 +0000 (0:00:00.757)       0:00:30.893 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Creating rabbitmq volume] *************************************\nFriday 13 March 2026  00:51:03 +0000 (0:00:01.760)       0:00:32.653 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Running RabbitMQ bootstrap container] *************************\nFriday 13 March 2026  00:51:04 +0000 (0:00:00.918)       0:00:33.571 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nFriday 13 March 2026  00:51:12 +0000 (0:00:07.647)       0:00:41.219 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nFriday 13 March 2026  00:51:12 +0000 (0:00:00.741)       0:00:41.961 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nFriday 13 March 2026  00:51:13 +0000 (0:00:00.909)       0:00:42.871 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nFriday 13 March 2026  00:51:13 +0000 (0:00:00.119)       0:00:42.990 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nFriday 13 March 2026  00:51:15 +0000 (0:00:02.008)       0:00:44.998 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nFriday 13 March 2026  00:53:08 +0000 (0:01:52.165)       0:02:37.164 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nFriday 13 March 2026  00:53:08 +0000 (0:00:00.612)       0:02:37.776 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nFriday 13 March 2026  00:53:08 +0000 (0:00:00.115)       0:02:37.892 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nFriday 13 March 2026  00:53:10 +0000 (0:00:01.472)       0:02:39.364 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nFriday 13 March 2026  00:53:19 +0000 (0:00:09.396)       0:02:48.761 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nFriday 13 March 2026  00:53:20 +0000 (0:00:00.567)       0:02:49.328 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nFriday 13 March 2026  00:53:20 +0000 (0:00:00.087)       0:02:49.416 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nFriday 13 March 2026  00:53:21 +0000 (0:00:01.422)       0:02:50.839 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nFriday 13 March 2026  00:53:30 +0000 (0:00:08.948)       0:02:59.787 ********** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nFriday 13 March 2026  00:53:31 +0000 (0:00:00.658)       0:03:00.445 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=10  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:53:34 +0000 (0:00:03.283)       0:03:03.729 ********** \n=============================================================================== \nrabbitmq : Waiting for rabbitmq to start ------------------------------ 130.51s\nrabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.65s\nrabbitmq : Restart rabbitmq container ----------------------------------- 4.90s\nCheck RabbitMQ service -------------------------------------------------- 4.26s\nrabbitmq : Enable all stable feature flags ------------------------------ 3.28s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.37s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 2.09s\nrabbitmq : Copying over advanced.config --------------------------------- 1.79s\nservice-check-containers : Include tasks -------------------------------- 1.76s\nrabbitmq : Copying over enabled_plugins --------------------------------- 1.63s\nrabbitmq : Get container facts ------------------------------------------ 1.53s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.51s\nrabbitmq : Copying over config.json files for services ------------------ 1.48s\nservice-cert-copy : rabbitmq | Copying over extra CA certificates ------- 1.45s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 1.37s\nservice-check-containers : rabbitmq | Check containers ------------------ 1.17s\nrabbitmq : Copying over definitions.json -------------------------------- 1.17s\nrabbitmq : Get new RabbitMQ version ------------------------------------- 1.00s\nrabbitmq : Ensuring config directories exist ---------------------------- 1.00s\nrabbitmq : Get current RabbitMQ version --------------------------------- 0.96s\n\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:51:14 +0000 (0:00:00.151)       0:00:00.151 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:51:15 +0000 (0:00:00.845)       0:00:00.997 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nFriday 13 March 2026  00:51:16 +0000 (0:00:00.929)       0:00:01.926 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nFriday 13 March 2026  00:51:17 +0000 (0:00:01.100)       0:00:03.026 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nFriday 13 March 2026  00:51:18 +0000 (0:00:01.171)       0:00:04.198 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nFriday 13 March 2026  00:51:19 +0000 (0:00:01.572)       0:00:05.771 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nFriday 13 March 2026  00:51:21 +0000 (0:00:01.243)       0:00:07.014 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_controller | Check containers] ************\nFriday 13 March 2026  00:51:22 +0000 (0:00:01.403)       0:00:08.418 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_controller | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:51:24 +0000 (0:00:01.514)       0:00:09.933 ********** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:51:24 +0000 (0:00:00.896)       0:00:10.829 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nFriday 13 March 2026  00:51:25 +0000 (0:00:01.021)       0:00:11.851 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nFriday 13 March 2026  00:51:28 +0000 (0:00:02.666)       0:00:14.518 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:22.823)       0:00:37.341 ********** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.062)       0:00:37.403 ********** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.081)       0:00:37.485 ********** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.061)       0:00:37.547 ********** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.063)       0:00:37.610 ********** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.074)       0:00:37.684 ********** \n\nRUNNING HANDLER [ovn-controller : Reload systemd config] ***********************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.080)       0:00:37.764 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nFriday 13 March 2026  00:51:54 +0000 (0:00:02.226)       0:00:39.991 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nFriday 13 March 2026  00:52:02 +0000 (0:00:08.256)       0:00:48.247 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nFriday 13 March 2026  00:52:03 +0000 (0:00:01.088)       0:00:49.336 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nFriday 13 March 2026  00:52:04 +0000 (0:00:01.260)       0:00:50.596 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nFriday 13 March 2026  00:52:05 +0000 (0:00:00.988)       0:00:51.585 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nFriday 13 March 2026  00:52:05 +0000 (0:00:00.286)       0:00:51.871 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nFriday 13 March 2026  00:52:06 +0000 (0:00:00.387)       0:00:52.259 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nFriday 13 March 2026  00:52:06 +0000 (0:00:00.297)       0:00:52.557 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nFriday 13 March 2026  00:52:06 +0000 (0:00:00.272)       0:00:52.829 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nFriday 13 March 2026  00:52:07 +0000 (0:00:00.275)       0:00:53.105 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nFriday 13 March 2026  00:52:07 +0000 (0:00:00.599)       0:00:53.704 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nFriday 13 March 2026  00:52:08 +0000 (0:00:00.375)       0:00:54.080 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nFriday 13 March 2026  00:52:08 +0000 (0:00:00.301)       0:00:54.381 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nFriday 13 March 2026  00:52:08 +0000 (0:00:00.301)       0:00:54.683 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nFriday 13 March 2026  00:52:09 +0000 (0:00:00.305)       0:00:54.988 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nFriday 13 March 2026  00:52:09 +0000 (0:00:00.421)       0:00:55.410 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nFriday 13 March 2026  00:52:09 +0000 (0:00:00.243)       0:00:55.654 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nFriday 13 March 2026  00:52:10 +0000 (0:00:00.258)       0:00:55.912 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nFriday 13 March 2026  00:52:10 +0000 (0:00:00.362)       0:00:56.274 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nFriday 13 March 2026  00:52:10 +0000 (0:00:00.444)       0:00:56.718 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nFriday 13 March 2026  00:52:11 +0000 (0:00:00.298)       0:00:57.017 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************\nFriday 13 March 2026  00:52:11 +0000 (0:00:00.541)       0:00:57.558 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************\nFriday 13 March 2026  00:52:12 +0000 (0:00:00.569)       0:00:58.128 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check NB cluster status] ****************************************\nFriday 13 March 2026  00:52:12 +0000 (0:00:00.387)       0:00:58.515 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check SB cluster status] ****************************************\nFriday 13 March 2026  00:52:12 +0000 (0:00:00.301)       0:00:58.816 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***\nFriday 13 March 2026  00:52:13 +0000 (0:00:00.282)       0:00:59.099 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***\nFriday 13 March 2026  00:52:13 +0000 (0:00:00.426)       0:00:59.525 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************\nFriday 13 March 2026  00:52:13 +0000 (0:00:00.271)       0:00:59.797 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************\nFriday 13 March 2026  00:52:14 +0000 (0:00:00.298)       0:01:00.095 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nFriday 13 March 2026  00:52:14 +0000 (0:00:00.277)       0:01:00.373 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nFriday 13 March 2026  00:52:17 +0000 (0:00:02.751)       0:01:03.125 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Ensure configuration for relays exists] *************************\nFriday 13 March 2026  00:52:23 +0000 (0:00:05.768)       0:01:08.893 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****\nFriday 13 March 2026  00:52:23 +0000 (0:00:00.691)       0:01:09.584 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for OVN relay services] **********\nFriday 13 March 2026  00:52:24 +0000 (0:00:00.662)       0:01:10.247 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Generate config files for OVN relay services] *******************\nFriday 13 March 2026  00:52:26 +0000 (0:00:01.828)       0:01:12.075 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers] ********************\nFriday 13 March 2026  00:52:28 +0000 (0:00:01.854)       0:01:13.929 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:52:32 +0000 (0:00:04.339)       0:01:18.269 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:52:32 +0000 (0:00:00.423)       0:01:18.693 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers with iteration] *****\nFriday 13 March 2026  00:52:35 +0000 (0:00:02.272)       0:01:20.965 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:52:35 +0000 (0:00:00.911)       0:01:21.876 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.580)       0:01:22.457 ********** \n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.072)       0:01:22.530 ********** \n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.062)       0:01:22.593 ********** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.052)       0:01:22.645 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nFriday 13 March 2026  00:52:44 +0000 (0:00:07.648)       0:01:30.294 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************\nFriday 13 March 2026  00:52:56 +0000 (0:00:11.780)       0:01:42.075 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nFriday 13 March 2026  00:53:10 +0000 (0:00:14.549)       0:01:56.624 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nFriday 13 March 2026  00:53:24 +0000 (0:00:13.437)       0:02:10.062 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nFriday 13 March 2026  00:53:24 +0000 (0:00:00.110)       0:02:10.172 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nFriday 13 March 2026  00:53:25 +0000 (0:00:00.753)       0:02:10.926 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nFriday 13 March 2026  00:53:25 +0000 (0:00:00.601)       0:02:11.527 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nFriday 13 March 2026  00:53:26 +0000 (0:00:01.008)       0:02:12.536 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nFriday 13 March 2026  00:53:27 +0000 (0:00:00.552)       0:02:13.088 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nFriday 13 March 2026  00:53:27 +0000 (0:00:00.740)       0:02:13.829 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************\nFriday 13 March 2026  00:53:28 +0000 (0:00:00.773)       0:02:14.602 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m\n\nTASK [ovn-db : Unset bootstrap args fact] **************************************\nFriday 13 March 2026  00:53:29 +0000 (0:00:01.092)       0:02:15.694 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nFriday 13 March 2026  00:53:30 +0000 (0:00:00.316)       0:02:16.011 ********** \n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nFriday 13 March 2026  00:53:34 +0000 (0:00:03.899)       0:02:19.910 ********** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [ovn-db : Ensure configuration for relays exists] *************************\nFriday 13 March 2026  00:53:39 +0000 (0:00:05.805)       0:02:25.716 ********** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****\nFriday 13 March 2026  00:53:40 +0000 (0:00:00.577)       0:02:26.294 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for OVN relay services] **********\nFriday 13 March 2026  00:53:41 +0000 (0:00:00.671)       0:02:26.965 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Generate config files for OVN relay services] *******************\nFriday 13 March 2026  00:53:42 +0000 (0:00:01.553)       0:02:28.519 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers] ********************\nFriday 13 March 2026  00:53:44 +0000 (0:00:01.750)       0:02:30.269 ********** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:53:49 +0000 (0:00:04.798)       0:02:35.068 ********** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:53:49 +0000 (0:00:00.302)       0:02:35.370 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-2, testbed-node-1 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : ovn_db | Check containers with iteration] *****\nFriday 13 March 2026  00:53:51 +0000 (0:00:01.851)       0:02:37.222 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m\n\nTASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:53:52 +0000 (0:00:01.011)       0:02:38.234 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:53:52 +0000 (0:00:00.495)       0:02:38.729 ********** \n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:53:52 +0000 (0:00:00.058)       0:02:38.787 ********** \n\nTASK [ovn-db : Flush handlers] *************************************************\nFriday 13 March 2026  00:53:52 +0000 (0:00:00.057)       0:02:38.845 ********** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nFriday 13 March 2026  00:53:53 +0000 (0:00:00.058)       0:02:38.904 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nFriday 13 March 2026  00:54:05 +0000 (0:00:12.287)       0:02:51.191 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************\nFriday 13 March 2026  00:54:18 +0000 (0:00:12.841)       0:03:04.032 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nFriday 13 March 2026  00:54:30 +0000 (0:00:12.650)       0:03:16.683 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nFriday 13 March 2026  00:54:30 +0000 (0:00:00.109)       0:03:16.792 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nFriday 13 March 2026  00:54:31 +0000 (0:00:00.744)       0:03:17.537 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nFriday 13 March 2026  00:54:32 +0000 (0:00:00.594)       0:03:18.132 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nFriday 13 March 2026  00:54:33 +0000 (0:00:00.926)       0:03:19.058 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nFriday 13 March 2026  00:54:33 +0000 (0:00:00.633)       0:03:19.692 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nFriday 13 March 2026  00:54:34 +0000 (0:00:00.723)       0:03:20.416 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************\nFriday 13 March 2026  00:54:35 +0000 (0:00:00.894)       0:03:21.310 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=65  \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=63  \u001b[0m \u001b[0;33mchanged=30  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=23  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=63  \u001b[0m \u001b[0;33mchanged=30  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=23  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:54:36 +0000 (0:00:01.294)       0:03:22.604 ********** \n=============================================================================== \novn-db : Restart ovn-sb-db-relay container ----------------------------- 27.20s\novn-db : Restart ovn-sb-db container ----------------------------------- 24.62s\novn-controller : Configure OVN in OVSDB -------------------------------- 22.82s\novn-db : Restart ovn-nb-db container ----------------------------------- 19.94s\novn-db : Restart ovn-northd container ---------------------------------- 13.44s\novn-controller : Restart ovn-controller container ----------------------- 8.26s\novn-db : Copying over config.json files for services -------------------- 5.81s\novn-db : Copying over config.json files for services -------------------- 5.77s\nservice-check-containers : ovn_db | Check containers -------------------- 4.80s\nservice-check-containers : ovn_db | Check containers -------------------- 4.34s\novn-db : Ensuring config directories exist ------------------------------ 3.90s\novn-db : Ensuring config directories exist ------------------------------ 2.75s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.67s\nservice-check-containers : Include tasks -------------------------------- 2.27s\novn-controller : Reload systemd config ---------------------------------- 2.23s\novn-db : Generate config files for OVN relay services ------------------- 1.85s\nservice-check-containers : Include tasks -------------------------------- 1.85s\novn-db : Copying over config.json files for OVN relay services ---------- 1.83s\novn-db : Generate config files for OVN relay services ------------------- 1.75s\novn-controller : Copying over config.json files for services ------------ 1.57s\n\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.376)       0:00:00.376 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:50:09 +0000 (0:00:00.419)       0:00:00.796 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nFriday 13 March 2026  00:50:10 +0000 (0:00:00.782)       0:00:01.579 ********** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nFriday 13 March 2026  00:50:11 +0000 (0:00:00.776)       0:00:02.355 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nFriday 13 March 2026  00:50:12 +0000 (0:00:00.921)       0:00:03.276 ********** \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nFriday 13 March 2026  00:50:13 +0000 (0:00:01.030)       0:00:04.307 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nFriday 13 March 2026  00:50:14 +0000 (0:00:00.829)       0:00:05.136 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nFriday 13 March 2026  00:50:17 +0000 (0:00:02.964)       0:00:08.101 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nFriday 13 March 2026  00:50:18 +0000 (0:00:01.138)       0:00:09.239 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nFriday 13 March 2026  00:50:19 +0000 (0:00:01.445)       0:00:10.685 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nFriday 13 March 2026  00:50:20 +0000 (0:00:01.173)       0:00:11.859 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nFriday 13 March 2026  00:50:22 +0000 (0:00:01.952)       0:00:13.811 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nFriday 13 March 2026  00:50:23 +0000 (0:00:01.001)       0:00:14.812 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nFriday 13 March 2026  00:50:26 +0000 (0:00:02.322)       0:00:17.135 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nFriday 13 March 2026  00:50:27 +0000 (0:00:01.306)       0:00:18.442 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nFriday 13 March 2026  00:50:30 +0000 (0:00:02.811)       0:00:21.253 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.911)       0:00:22.165 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nFriday 13 March 2026  00:50:35 +0000 (0:00:04.355)       0:00:26.520 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nFriday 13 March 2026  00:50:40 +0000 (0:00:04.744)       0:00:31.265 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nFriday 13 March 2026  00:50:42 +0000 (0:00:02.533)       0:00:33.798 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nFriday 13 March 2026  00:50:46 +0000 (0:00:03.561)       0:00:37.360 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nFriday 13 March 2026  00:50:46 +0000 (0:00:00.626)       0:00:37.986 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nFriday 13 March 2026  00:50:49 +0000 (0:00:02.865)       0:00:40.851 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nFriday 13 March 2026  00:50:52 +0000 (0:00:02.256)       0:00:43.108 ********** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nFriday 13 March 2026  00:50:52 +0000 (0:00:00.742)       0:00:43.850 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nFriday 13 March 2026  00:50:54 +0000 (0:00:01.713)       0:00:45.564 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql-cert.pem] ***************************\nFriday 13 March 2026  00:50:56 +0000 (0:00:02.235)       0:00:47.800 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql-key.pem] ****************************\nFriday 13 March 2026  00:50:57 +0000 (0:00:00.320)       0:00:48.120 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nFriday 13 March 2026  00:50:57 +0000 (0:00:00.532)       0:00:48.653 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:51:00 +0000 (0:00:02.944)       0:00:51.597 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nFriday 13 March 2026  00:51:03 +0000 (0:00:02.685)       0:00:54.282 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nFriday 13 March 2026  00:51:04 +0000 (0:00:01.113)       0:00:55.396 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nFriday 13 March 2026  00:51:05 +0000 (0:00:01.319)       0:00:56.715 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nFriday 13 March 2026  00:51:07 +0000 (0:00:01.475)       0:00:58.190 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Check containers] **************\nFriday 13 March 2026  00:51:08 +0000 (0:00:00.848)       0:00:59.039 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:51:10 +0000 (0:00:02.467)       0:01:01.507 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:51:10 +0000 (0:00:00.323)       0:01:01.830 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nFriday 13 March 2026  00:51:12 +0000 (0:00:01.272)       0:01:03.103 ********** \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nFriday 13 March 2026  00:51:12 +0000 (0:00:00.596)       0:01:03.699 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nFriday 13 March 2026  00:51:16 +0000 (0:00:03.953)       0:01:07.653 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nFriday 13 March 2026  00:51:17 +0000 (0:00:00.577)       0:01:08.231 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nFriday 13 March 2026  00:51:18 +0000 (0:00:00.821)       0:01:09.052 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nFriday 13 March 2026  00:51:19 +0000 (0:00:01.377)       0:01:10.430 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nFriday 13 March 2026  00:51:21 +0000 (0:00:01.808)       0:01:12.238 ********** \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nFriday 13 March 2026  00:51:21 +0000 (0:00:00.608)       0:01:12.847 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nFriday 13 March 2026  00:51:25 +0000 (0:00:04.133)       0:01:16.980 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nFriday 13 March 2026  00:51:26 +0000 (0:00:00.716)       0:01:17.697 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nFriday 13 March 2026  00:51:27 +0000 (0:00:00.889)       0:01:18.586 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nFriday 13 March 2026  00:51:28 +0000 (0:00:01.267)       0:01:19.854 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nFriday 13 March 2026  00:51:32 +0000 (0:00:03.311)       0:01:23.165 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nFriday 13 March 2026  00:51:32 +0000 (0:00:00.273)       0:01:23.439 ********** \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nFriday 13 March 2026  00:51:33 +0000 (0:00:00.761)       0:01:24.201 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nFriday 13 March 2026  00:51:35 +0000 (0:00:02.677)       0:01:26.878 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nFriday 13 March 2026  00:51:37 +0000 (0:00:01.433)       0:01:28.312 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nFriday 13 March 2026  00:51:38 +0000 (0:00:01.633)       0:01:29.946 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nFriday 13 March 2026  00:51:39 +0000 (0:00:00.430)       0:01:30.376 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nFriday 13 March 2026  00:51:40 +0000 (0:00:01.034)       0:01:31.410 ********** \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nFriday 13 March 2026  00:51:41 +0000 (0:00:00.778)       0:01:32.189 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nFriday 13 March 2026  00:51:44 +0000 (0:00:03.261)       0:01:35.451 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nFriday 13 March 2026  00:51:45 +0000 (0:00:00.948)       0:01:36.400 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nFriday 13 March 2026  00:51:46 +0000 (0:00:01.329)       0:01:37.729 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nFriday 13 March 2026  00:51:48 +0000 (0:00:01.524)       0:01:39.254 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nFriday 13 March 2026  00:51:50 +0000 (0:00:01.929)       0:01:41.184 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nFriday 13 March 2026  00:51:50 +0000 (0:00:00.232)       0:01:41.416 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nFriday 13 March 2026  00:51:50 +0000 (0:00:00.230)       0:01:41.647 ********** \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nFriday 13 March 2026  00:51:51 +0000 (0:00:00.718)       0:01:42.365 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nFriday 13 March 2026  00:51:55 +0000 (0:00:03.877)       0:01:46.243 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nFriday 13 March 2026  00:51:56 +0000 (0:00:00.976)       0:01:47.220 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nFriday 13 March 2026  00:51:57 +0000 (0:00:01.297)       0:01:48.518 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nFriday 13 March 2026  00:51:58 +0000 (0:00:01.163)       0:01:49.682 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nFriday 13 March 2026  00:52:00 +0000 (0:00:01.790)       0:01:51.473 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nFriday 13 March 2026  00:52:00 +0000 (0:00:00.273)       0:01:51.746 ********** \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nFriday 13 March 2026  00:52:01 +0000 (0:00:00.841)       0:01:52.587 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nFriday 13 March 2026  00:52:06 +0000 (0:00:05.205)       0:01:57.793 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nFriday 13 March 2026  00:52:09 +0000 (0:00:02.656)       0:02:00.449 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nFriday 13 March 2026  00:52:12 +0000 (0:00:02.982)       0:02:03.432 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nFriday 13 March 2026  00:52:13 +0000 (0:00:01.317)       0:02:04.749 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nFriday 13 March 2026  00:52:15 +0000 (0:00:01.889)       0:02:06.639 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nFriday 13 March 2026  00:52:15 +0000 (0:00:00.305)       0:02:06.944 ********** \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nFriday 13 March 2026  00:52:16 +0000 (0:00:00.991)       0:02:07.936 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nFriday 13 March 2026  00:52:20 +0000 (0:00:03.716)       0:02:11.652 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nFriday 13 March 2026  00:52:21 +0000 (0:00:00.380)       0:02:12.033 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nFriday 13 March 2026  00:52:21 +0000 (0:00:00.614)       0:02:12.648 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nFriday 13 March 2026  00:52:23 +0000 (0:00:01.492)       0:02:14.141 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nFriday 13 March 2026  00:52:25 +0000 (0:00:02.372)       0:02:16.513 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nFriday 13 March 2026  00:52:25 +0000 (0:00:00.326)       0:02:16.839 ********** \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nFriday 13 March 2026  00:52:26 +0000 (0:00:00.917)       0:02:17.756 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nFriday 13 March 2026  00:52:31 +0000 (0:00:04.757)       0:02:22.514 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nFriday 13 March 2026  00:52:32 +0000 (0:00:00.584)       0:02:23.099 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nFriday 13 March 2026  00:52:33 +0000 (0:00:00.945)       0:02:24.044 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nFriday 13 March 2026  00:52:34 +0000 (0:00:01.763)       0:02:25.808 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:01.767)       0:02:27.576 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.242)       0:02:27.819 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nFriday 13 March 2026  00:52:37 +0000 (0:00:00.259)       0:02:28.078 ********** \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nFriday 13 March 2026  00:52:38 +0000 (0:00:01.019)       0:02:29.097 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nFriday 13 March 2026  00:52:41 +0000 (0:00:03.857)       0:02:32.955 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nFriday 13 March 2026  00:52:42 +0000 (0:00:00.684)       0:02:33.640 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nFriday 13 March 2026  00:52:43 +0000 (0:00:00.868)       0:02:34.508 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nFriday 13 March 2026  00:52:44 +0000 (0:00:01.159)       0:02:35.667 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nFriday 13 March 2026  00:52:46 +0000 (0:00:01.845)       0:02:37.513 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nFriday 13 March 2026  00:52:46 +0000 (0:00:00.319)       0:02:37.833 ********** \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nFriday 13 March 2026  00:52:48 +0000 (0:00:01.420)       0:02:39.254 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nFriday 13 March 2026  00:52:52 +0000 (0:00:03.861)       0:02:43.115 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nFriday 13 March 2026  00:52:52 +0000 (0:00:00.560)       0:02:43.676 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nFriday 13 March 2026  00:52:53 +0000 (0:00:00.801)       0:02:44.477 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nFriday 13 March 2026  00:52:54 +0000 (0:00:01.274)       0:02:45.751 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nFriday 13 March 2026  00:52:56 +0000 (0:00:01.932)       0:02:47.684 ********** \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nFriday 13 March 2026  00:52:57 +0000 (0:00:01.014)       0:02:48.698 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nFriday 13 March 2026  00:53:02 +0000 (0:00:05.110)       0:02:53.808 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nFriday 13 March 2026  00:53:03 +0000 (0:00:01.047)       0:02:54.856 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nFriday 13 March 2026  00:53:04 +0000 (0:00:00.763)       0:02:55.620 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nFriday 13 March 2026  00:53:05 +0000 (0:00:01.219)       0:02:56.839 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nFriday 13 March 2026  00:53:07 +0000 (0:00:01.788)       0:02:58.627 ********** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nFriday 13 March 2026  00:53:08 +0000 (0:00:01.102)       0:02:59.730 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nFriday 13 March 2026  00:53:11 +0000 (0:00:03.167)       0:03:02.897 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nFriday 13 March 2026  00:53:15 +0000 (0:00:03.348)       0:03:06.246 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nFriday 13 March 2026  00:53:17 +0000 (0:00:02.755)       0:03:09.002 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nFriday 13 March 2026  00:53:19 +0000 (0:00:01.864)       0:03:10.867 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nFriday 13 March 2026  00:53:21 +0000 (0:00:01.621)       0:03:12.488 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nFriday 13 March 2026  00:53:22 +0000 (0:00:01.258)       0:03:13.746 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nFriday 13 March 2026  00:53:22 +0000 (0:00:00.261)       0:03:14.008 ********** \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nFriday 13 March 2026  00:53:23 +0000 (0:00:00.933)       0:03:14.941 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nFriday 13 March 2026  00:53:25 +0000 (0:00:01.332)       0:03:16.273 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nFriday 13 March 2026  00:53:25 +0000 (0:00:00.378)       0:03:16.652 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nFriday 13 March 2026  00:53:26 +0000 (0:00:00.849)       0:03:17.502 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nFriday 13 March 2026  00:53:26 +0000 (0:00:00.418)       0:03:17.920 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nFriday 13 March 2026  00:53:28 +0000 (0:00:01.216)       0:03:19.137 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nFriday 13 March 2026  00:53:28 +0000 (0:00:00.300)       0:03:19.437 ********** \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nFriday 13 March 2026  00:53:29 +0000 (0:00:01.397)       0:03:20.835 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nFriday 13 March 2026  00:53:35 +0000 (0:00:05.803)       0:03:26.638 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nFriday 13 March 2026  00:53:37 +0000 (0:00:02.158)       0:03:28.796 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nFriday 13 March 2026  00:53:39 +0000 (0:00:01.506)       0:03:30.303 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nFriday 13 March 2026  00:53:40 +0000 (0:00:01.208)       0:03:31.512 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nFriday 13 March 2026  00:53:42 +0000 (0:00:02.130)       0:03:33.643 ********** \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nFriday 13 March 2026  00:53:44 +0000 (0:00:01.379)       0:03:35.022 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nFriday 13 March 2026  00:53:47 +0000 (0:00:03.946)       0:03:38.969 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nFriday 13 March 2026  00:53:48 +0000 (0:00:00.430)       0:03:39.399 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nFriday 13 March 2026  00:53:49 +0000 (0:00:00.885)       0:03:40.285 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nFriday 13 March 2026  00:53:50 +0000 (0:00:01.130)       0:03:41.415 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nFriday 13 March 2026  00:53:52 +0000 (0:00:01.857)       0:03:43.273 ********** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nFriday 13 March 2026  00:53:53 +0000 (0:00:01.168)       0:03:44.442 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nFriday 13 March 2026  00:54:00 +0000 (0:00:06.585)       0:03:51.028 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nFriday 13 March 2026  00:54:00 +0000 (0:00:00.680)       0:03:51.708 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nFriday 13 March 2026  00:54:01 +0000 (0:00:00.813)       0:03:52.521 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nFriday 13 March 2026  00:54:02 +0000 (0:00:01.306)       0:03:53.828 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nFriday 13 March 2026  00:54:04 +0000 (0:00:01.893)       0:03:55.722 ********** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nFriday 13 March 2026  00:54:06 +0000 (0:00:01.395)       0:03:57.117 ********** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nFriday 13 March 2026  00:54:07 +0000 (0:00:01.488)       0:03:58.606 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nFriday 13 March 2026  00:54:12 +0000 (0:00:04.675)       0:04:03.282 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nFriday 13 March 2026  00:54:13 +0000 (0:00:01.107)       0:04:04.390 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nFriday 13 March 2026  00:54:14 +0000 (0:00:01.559)       0:04:05.949 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nFriday 13 March 2026  00:54:17 +0000 (0:00:02.171)       0:04:08.120 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nFriday 13 March 2026  00:54:20 +0000 (0:00:03.231)       0:04:11.352 ********** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-2, testbed-node-0, testbed-node-1 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nFriday 13 March 2026  00:54:21 +0000 (0:00:00.860)       0:04:12.213 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nFriday 13 March 2026  00:54:22 +0000 (0:00:01.679)       0:04:13.893 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nFriday 13 March 2026  00:54:24 +0000 (0:00:01.938)       0:04:15.831 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nFriday 13 March 2026  00:54:25 +0000 (0:00:01.038)       0:04:16.870 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nFriday 13 March 2026  00:54:27 +0000 (0:00:01.980)       0:04:18.850 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nFriday 13 March 2026  00:54:30 +0000 (0:00:02.660)       0:04:21.511 ********** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nFriday 13 March 2026  00:54:31 +0000 (0:00:00.756)       0:04:22.267 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nFriday 13 March 2026  00:54:32 +0000 (0:00:01.390)       0:04:23.657 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nFriday 13 March 2026  00:54:33 +0000 (0:00:01.218)       0:04:24.876 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nFriday 13 March 2026  00:54:35 +0000 (0:00:01.458)       0:04:26.334 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nFriday 13 March 2026  00:54:37 +0000 (0:00:02.401)       0:04:28.736 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nFriday 13 March 2026  00:54:40 +0000 (0:00:03.110)       0:04:31.847 ********** \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nFriday 13 March 2026  00:54:42 +0000 (0:00:01.276)       0:04:33.123 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nFriday 13 March 2026  00:54:45 +0000 (0:00:03.645)       0:04:36.769 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nFriday 13 March 2026  00:54:46 +0000 (0:00:00.961)       0:04:37.730 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nFriday 13 March 2026  00:54:47 +0000 (0:00:00.808)       0:04:38.539 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nFriday 13 March 2026  00:54:48 +0000 (0:00:01.268)       0:04:39.808 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nFriday 13 March 2026  00:54:50 +0000 (0:00:01.919)       0:04:41.727 ********** \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nFriday 13 March 2026  00:54:52 +0000 (0:00:01.403)       0:04:43.130 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nFriday 13 March 2026  00:54:56 +0000 (0:00:04.654)       0:04:47.785 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nFriday 13 March 2026  00:54:57 +0000 (0:00:00.828)       0:04:48.614 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nFriday 13 March 2026  00:54:58 +0000 (0:00:01.103)       0:04:49.717 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nFriday 13 March 2026  00:54:59 +0000 (0:00:00.383)       0:04:50.101 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nFriday 13 March 2026  00:55:00 +0000 (0:00:01.277)       0:04:51.378 ********** \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nFriday 13 March 2026  00:55:02 +0000 (0:00:01.689)       0:04:53.068 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nFriday 13 March 2026  00:55:06 +0000 (0:00:04.417)       0:04:57.485 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nFriday 13 March 2026  00:55:07 +0000 (0:00:00.825)       0:04:58.310 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nFriday 13 March 2026  00:55:08 +0000 (0:00:01.258)       0:04:59.569 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nFriday 13 March 2026  00:55:09 +0000 (0:00:00.467)       0:05:00.036 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nFriday 13 March 2026  00:55:10 +0000 (0:00:01.310)       0:05:01.347 ********** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nFriday 13 March 2026  00:55:11 +0000 (0:00:01.410)       0:05:02.757 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nFriday 13 March 2026  00:55:14 +0000 (0:00:02.433)       0:05:05.190 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nFriday 13 March 2026  00:55:14 +0000 (0:00:00.464)       0:05:05.655 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nFriday 13 March 2026  00:55:15 +0000 (0:00:00.620)       0:05:06.276 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nFriday 13 March 2026  00:55:16 +0000 (0:00:00.847)       0:05:07.124 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nFriday 13 March 2026  00:55:17 +0000 (0:00:01.297)       0:05:08.422 ********** \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nFriday 13 March 2026  00:55:18 +0000 (0:00:01.499)       0:05:09.921 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nFriday 13 March 2026  00:55:24 +0000 (0:00:05.922)       0:05:15.844 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nFriday 13 March 2026  00:55:25 +0000 (0:00:01.134)       0:05:16.978 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nFriday 13 March 2026  00:55:26 +0000 (0:00:00.942)       0:05:17.920 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nFriday 13 March 2026  00:55:28 +0000 (0:00:01.132)       0:05:19.053 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nFriday 13 March 2026  00:55:30 +0000 (0:00:01.977)       0:05:21.030 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nFriday 13 March 2026  00:55:30 +0000 (0:00:00.603)       0:05:21.634 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nFriday 13 March 2026  00:55:30 +0000 (0:00:00.332)       0:05:21.966 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nFriday 13 March 2026  00:55:31 +0000 (0:00:00.298)       0:05:22.265 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nFriday 13 March 2026  00:55:31 +0000 (0:00:00.312)       0:05:22.577 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : loadbalancer] *********************************************\nFriday 13 March 2026  00:55:31 +0000 (0:00:00.299)       0:05:22.877 ********** \n\u001b[0;36mincluded: loadbalancer for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Check containers] **************\nFriday 13 March 2026  00:55:33 +0000 (0:00:01.781)       0:05:24.658 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:55:36 +0000 (0:00:02.571)       0:05:27.230 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:55:36 +0000 (0:00:00.663)       0:05:27.893 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nFriday 13 March 2026  00:55:38 +0000 (0:00:01.529)       0:05:29.423 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nFriday 13 March 2026  00:55:39 +0000 (0:00:00.675)       0:05:30.099 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nFriday 13 March 2026  00:55:39 +0000 (0:00:00.310)       0:05:30.410 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nFriday 13 March 2026  00:55:40 +0000 (0:00:00.744)       0:05:31.154 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nFriday 13 March 2026  00:55:41 +0000 (0:00:01.073)       0:05:32.227 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nFriday 13 March 2026  00:55:42 +0000 (0:00:00.850)       0:05:33.078 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nFriday 13 March 2026  00:55:50 +0000 (0:00:08.898)       0:05:41.976 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nFriday 13 March 2026  00:55:51 +0000 (0:00:00.870)       0:05:42.847 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nFriday 13 March 2026  00:55:59 +0000 (0:00:08.109)       0:05:50.957 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nFriday 13 March 2026  00:56:03 +0000 (0:00:03.106)       0:05:54.063 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nFriday 13 March 2026  00:56:07 +0000 (0:00:04.401)       0:05:58.464 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nFriday 13 March 2026  00:56:07 +0000 (0:00:00.353)       0:05:58.818 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nFriday 13 March 2026  00:56:08 +0000 (0:00:00.352)       0:05:59.171 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nFriday 13 March 2026  00:56:08 +0000 (0:00:00.681)       0:05:59.852 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nFriday 13 March 2026  00:56:09 +0000 (0:00:00.357)       0:06:00.210 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nFriday 13 March 2026  00:56:09 +0000 (0:00:00.365)       0:06:00.575 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nFriday 13 March 2026  00:56:09 +0000 (0:00:00.344)       0:06:00.920 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nFriday 13 March 2026  00:56:15 +0000 (0:00:05.096)       0:06:06.016 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=126 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=126 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:56:15 +0000 (0:00:00.918)       0:06:06.935 ********** \n=============================================================================== \nloadbalancer : Start backup haproxy container --------------------------- 8.90s\nloadbalancer : Start backup proxysql container -------------------------- 8.11s\nhaproxy-config : Copying over nova haproxy config ----------------------- 6.59s\nhaproxy-config : Copying over skyline haproxy config -------------------- 5.92s\nhaproxy-config : Copying over neutron haproxy config -------------------- 5.80s\nhaproxy-config : Copying over glance haproxy config --------------------- 5.21s\nhaproxy-config : Copying over manila haproxy config --------------------- 5.11s\nloadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.10s\nhaproxy-config : Copying over horizon haproxy config -------------------- 4.76s\nloadbalancer : Copying over config.json files for services -------------- 4.74s\nhaproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.68s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 4.65s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 4.42s\nloadbalancer : Start backup keepalived container ------------------------ 4.40s\nloadbalancer : Copying checks for services which are enabled ------------ 4.36s\nhaproxy-config : Copying over barbican haproxy config ------------------- 4.13s\nhaproxy-config : Copying over aodh haproxy config ----------------------- 3.95s\nhaproxy-config : Copying over placement haproxy config ------------------ 3.95s\nhaproxy-config : Copying over designate haproxy config ------------------ 3.88s\nhaproxy-config : Copying over magnum haproxy config --------------------- 3.86s\n\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Prepare deployment of Ceph services] *************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nFriday 13 March 2026  00:47:35 +0000 (0:00:00.660)       0:00:00.660 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nFriday 13 March 2026  00:47:36 +0000 (0:00:01.083)       0:00:01.743 ********** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nFriday 13 March 2026  00:47:37 +0000 (0:00:01.677)       0:00:03.421 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nFriday 13 March 2026  00:47:38 +0000 (0:00:01.040)       0:00:04.461 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nFriday 13 March 2026  00:47:39 +0000 (0:00:01.044)       0:00:05.506 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nFriday 13 March 2026  00:47:40 +0000 (0:00:00.851)       0:00:06.358 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nFriday 13 March 2026  00:47:41 +0000 (0:00:00.633)       0:00:06.992 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nFriday 13 March 2026  00:47:42 +0000 (0:00:01.136)       0:00:08.128 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nFriday 13 March 2026  00:47:43 +0000 (0:00:00.777)       0:00:08.905 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nFriday 13 March 2026  00:47:44 +0000 (0:00:00.906)       0:00:09.812 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nFriday 13 March 2026  00:47:44 +0000 (0:00:00.471)       0:00:10.283 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nFriday 13 March 2026  00:47:46 +0000 (0:00:01.951)       0:00:12.235 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nFriday 13 March 2026  00:47:49 +0000 (0:00:02.440)       0:00:14.675 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nFriday 13 March 2026  00:47:49 +0000 (0:00:00.584)       0:00:15.259 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nFriday 13 March 2026  00:47:50 +0000 (0:00:00.749)       0:00:16.009 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nFriday 13 March 2026  00:47:51 +0000 (0:00:00.626)       0:00:16.635 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-13 00:47:47.390160', 'end': '2026-03-13 00:47:47.452558', 'delta': '0:00:00.062398', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-13 00:47:48.417953', 'end': '2026-03-13 00:47:48.478328', 'delta': '0:00:00.060375', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-13 00:47:48.949771', 'end': '2026-03-13 00:47:49.021265', 'delta': '0:00:00.071494', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nFriday 13 March 2026  00:47:51 +0000 (0:00:00.297)       0:00:16.933 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nFriday 13 March 2026  00:47:53 +0000 (0:00:01.855)       0:00:18.788 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nFriday 13 March 2026  00:47:53 +0000 (0:00:00.576)       0:00:19.364 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nFriday 13 March 2026  00:47:54 +0000 (0:00:01.095)       0:00:20.459 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nFriday 13 March 2026  00:47:56 +0000 (0:00:01.814)       0:00:22.274 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nFriday 13 March 2026  00:47:58 +0000 (0:00:01.456)       0:00:23.731 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nFriday 13 March 2026  00:47:58 +0000 (0:00:00.266)       0:00:23.997 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nFriday 13 March 2026  00:47:58 +0000 (0:00:00.514)       0:00:24.512 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nFriday 13 March 2026  00:48:00 +0000 (0:00:01.542)       0:00:26.055 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nFriday 13 March 2026  00:48:02 +0000 (0:00:02.334)       0:00:28.389 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nFriday 13 March 2026  00:48:04 +0000 (0:00:01.569)       0:00:29.959 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nFriday 13 March 2026  00:48:06 +0000 (0:00:01.652)       0:00:31.611 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nFriday 13 March 2026  00:48:07 +0000 (0:00:01.369)       0:00:32.982 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nFriday 13 March 2026  00:48:08 +0000 (0:00:01.261)       0:00:34.243 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nFriday 13 March 2026  00:48:09 +0000 (0:00:00.687)       0:00:34.930 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-39-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nFriday 13 March 2026  00:48:10 +0000 (0:00:01.447)       0:00:36.377 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-39-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nFriday 13 March 2026  00:48:11 +0000 (0:00:00.849)       0:00:37.227 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nFriday 13 March 2026  00:48:12 +0000 (0:00:01.177)       0:00:38.404 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nFriday 13 March 2026  00:48:13 +0000 (0:00:00.622)       0:00:39.027 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nFriday 13 March 2026  00:48:14 +0000 (0:00:00.760)       0:00:39.787 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nFriday 13 March 2026  00:48:15 +0000 (0:00:00.778)       0:00:40.566 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nFriday 13 March 2026  00:48:16 +0000 (0:00:01.200)       0:00:41.766 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nFriday 13 March 2026  00:48:17 +0000 (0:00:01.089)       0:00:42.856 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nFriday 13 March 2026  00:48:20 +0000 (0:00:02.819)       0:00:45.676 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nFriday 13 March 2026  00:48:21 +0000 (0:00:01.066)       0:00:46.742 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nFriday 13 March 2026  00:48:22 +0000 (0:00:01.144)       0:00:47.887 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nFriday 13 March 2026  00:48:22 +0000 (0:00:00.398)       0:00:48.286 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nFriday 13 March 2026  00:48:23 +0000 (0:00:00.295)       0:00:48.582 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nFriday 13 March 2026  00:48:23 +0000 (0:00:00.416)       0:00:48.999 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nFriday 13 March 2026  00:48:23 +0000 (0:00:00.376)       0:00:49.375 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nFriday 13 March 2026  00:48:24 +0000 (0:00:00.355)       0:00:49.731 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nFriday 13 March 2026  00:48:24 +0000 (0:00:00.342)       0:00:50.073 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nFriday 13 March 2026  00:48:24 +0000 (0:00:00.371)       0:00:50.445 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nFriday 13 March 2026  00:48:25 +0000 (0:00:00.285)       0:00:50.730 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nFriday 13 March 2026  00:48:25 +0000 (0:00:00.688)       0:00:51.418 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nFriday 13 March 2026  00:48:26 +0000 (0:00:00.750)       0:00:52.169 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:48:28 +0000 (0:00:01.640)       0:00:53.809 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:48:29 +0000 (0:00:01.091)       0:00:54.900 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:48:30 +0000 (0:00:01.330)       0:00:56.231 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:48:32 +0000 (0:00:01.302)       0:00:57.533 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:48:32 +0000 (0:00:00.858)       0:00:58.392 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:48:33 +0000 (0:00:00.758)       0:00:59.150 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:48:34 +0000 (0:00:00.853)       0:01:00.004 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:48:36 +0000 (0:00:01.611)       0:01:01.615 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:48:37 +0000 (0:00:00.931)       0:01:02.547 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:48:37 +0000 (0:00:00.905)       0:01:03.453 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:48:39 +0000 (0:00:01.131)       0:01:04.584 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:48:40 +0000 (0:00:01.541)       0:01:06.126 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:48:41 +0000 (0:00:00.946)       0:01:07.072 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:48:42 +0000 (0:00:01.180)       0:01:08.253 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:48:44 +0000 (0:00:01.336)       0:01:09.590 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:48:45 +0000 (0:00:01.338)       0:01:10.928 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:48:46 +0000 (0:00:00.897)       0:01:11.826 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:48:47 +0000 (0:00:01.274)       0:01:13.101 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:48:48 +0000 (0:00:00.796)       0:01:13.898 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:48:49 +0000 (0:00:00.842)       0:01:14.741 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:48:50 +0000 (0:00:01.194)       0:01:15.935 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nFriday 13 March 2026  00:48:52 +0000 (0:00:02.493)       0:01:18.429 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nFriday 13 March 2026  00:48:55 +0000 (0:00:02.618)       0:01:21.047 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nFriday 13 March 2026  00:48:59 +0000 (0:00:03.483)       0:01:24.531 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nFriday 13 March 2026  00:49:00 +0000 (0:00:01.174)       0:01:25.705 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nFriday 13 March 2026  00:49:01 +0000 (0:00:00.923)       0:01:26.628 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nFriday 13 March 2026  00:49:02 +0000 (0:00:00.937)       0:01:27.566 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nFriday 13 March 2026  00:49:03 +0000 (0:00:01.475)       0:01:29.041 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nFriday 13 March 2026  00:49:04 +0000 (0:00:01.431)       0:01:30.472 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nFriday 13 March 2026  00:49:05 +0000 (0:00:00.611)       0:01:31.084 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nFriday 13 March 2026  00:49:06 +0000 (0:00:00.803)       0:01:31.888 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nFriday 13 March 2026  00:49:06 +0000 (0:00:00.584)       0:01:32.472 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nFriday 13 March 2026  00:49:08 +0000 (0:00:01.221)       0:01:33.694 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nFriday 13 March 2026  00:49:50 +0000 (0:00:42.045)       0:02:15.739 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nFriday 13 March 2026  00:49:50 +0000 (0:00:00.580)       0:02:16.319 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nFriday 13 March 2026  00:49:51 +0000 (0:00:00.847)       0:02:17.166 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nFriday 13 March 2026  00:49:51 +0000 (0:00:00.161)       0:02:17.328 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nFriday 13 March 2026  00:49:52 +0000 (0:00:00.580)       0:02:17.908 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nFriday 13 March 2026  00:49:53 +0000 (0:00:00.658)       0:02:18.566 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nFriday 13 March 2026  00:49:53 +0000 (0:00:00.721)       0:02:19.288 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nFriday 13 March 2026  00:49:57 +0000 (0:00:03.520)       0:02:22.809 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nFriday 13 March 2026  00:49:57 +0000 (0:00:00.605)       0:02:23.415 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nFriday 13 March 2026  00:49:58 +0000 (0:00:01.034)       0:02:24.449 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nFriday 13 March 2026  00:49:59 +0000 (0:00:00.654)       0:02:25.104 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nFriday 13 March 2026  00:50:00 +0000 (0:00:00.497)       0:02:25.602 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nFriday 13 March 2026  00:50:00 +0000 (0:00:00.656)       0:02:26.258 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nFriday 13 March 2026  00:50:01 +0000 (0:00:00.520)       0:02:26.779 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nFriday 13 March 2026  00:50:01 +0000 (0:00:00.639)       0:02:27.418 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nFriday 13 March 2026  00:50:02 +0000 (0:00:00.536)       0:02:27.954 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nFriday 13 March 2026  00:50:03 +0000 (0:00:00.799)       0:02:28.754 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nFriday 13 March 2026  00:50:03 +0000 (0:00:00.612)       0:02:29.366 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nFriday 13 March 2026  00:50:05 +0000 (0:00:01.256)       0:02:30.623 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nFriday 13 March 2026  00:50:06 +0000 (0:00:01.327)       0:02:31.950 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nFriday 13 March 2026  00:50:12 +0000 (0:00:06.457)       0:02:38.407 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nFriday 13 March 2026  00:50:13 +0000 (0:00:00.878)       0:02:39.286 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nFriday 13 March 2026  00:50:14 +0000 (0:00:00.918)       0:02:40.205 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nFriday 13 March 2026  00:50:16 +0000 (0:00:01.629)       0:02:41.834 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nFriday 13 March 2026  00:50:16 +0000 (0:00:00.681)       0:02:42.515 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nFriday 13 March 2026  00:50:17 +0000 (0:00:00.903)       0:02:43.419 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nFriday 13 March 2026  00:50:18 +0000 (0:00:00.565)       0:02:43.985 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nFriday 13 March 2026  00:50:19 +0000 (0:00:00.741)       0:02:44.726 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nFriday 13 March 2026  00:50:19 +0000 (0:00:00.585)       0:02:45.312 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nFriday 13 March 2026  00:50:20 +0000 (0:00:00.782)       0:02:46.095 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nFriday 13 March 2026  00:50:21 +0000 (0:00:00.797)       0:02:46.893 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nFriday 13 March 2026  00:50:22 +0000 (0:00:01.065)       0:02:47.958 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nFriday 13 March 2026  00:50:25 +0000 (0:00:02.929)       0:02:50.887 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nFriday 13 March 2026  00:50:26 +0000 (0:00:00.879)       0:02:51.767 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nFriday 13 March 2026  00:50:26 +0000 (0:00:00.739)       0:02:52.507 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nFriday 13 March 2026  00:50:27 +0000 (0:00:00.748)       0:02:53.255 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nFriday 13 March 2026  00:50:28 +0000 (0:00:00.728)       0:02:53.984 ********** \n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nFriday 13 March 2026  00:50:29 +0000 (0:00:01.233)       0:02:55.218 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nFriday 13 March 2026  00:50:30 +0000 (0:00:00.438)       0:02:55.656 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nFriday 13 March 2026  00:50:30 +0000 (0:00:00.612)       0:02:56.269 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nFriday 13 March 2026  00:50:31 +0000 (0:00:00.579)       0:02:56.848 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nFriday 13 March 2026  00:50:32 +0000 (0:00:00.715)       0:02:57.564 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nFriday 13 March 2026  00:50:32 +0000 (0:00:00.723)       0:02:58.288 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nFriday 13 March 2026  00:50:34 +0000 (0:00:01.374)       0:02:59.662 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nFriday 13 March 2026  00:50:34 +0000 (0:00:00.452)       0:03:00.114 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nFriday 13 March 2026  00:50:34 +0000 (0:00:00.352)       0:03:00.467 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.339)       0:03:00.807 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nFriday 13 March 2026  00:50:35 +0000 (0:00:00.562)       0:03:01.370 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nFriday 13 March 2026  00:50:37 +0000 (0:00:01.808)       0:03:03.178 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:50:40 +0000 (0:00:02.425)       0:03:05.603 ********** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nFriday 13 March 2026  00:50:41 +0000 (0:00:01.377)       0:03:06.981 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nFriday 13 March 2026  00:50:42 +0000 (0:00:00.868)       0:03:07.849 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nFriday 13 March 2026  00:50:42 +0000 (0:00:00.233)       0:03:08.082 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nFriday 13 March 2026  00:50:43 +0000 (0:00:01.197)       0:03:09.280 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nFriday 13 March 2026  00:50:44 +0000 (0:00:00.574)       0:03:09.854 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nFriday 13 March 2026  00:50:44 +0000 (0:00:00.346)       0:03:10.201 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nFriday 13 March 2026  00:50:45 +0000 (0:00:00.984)       0:03:11.185 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nFriday 13 March 2026  00:50:45 +0000 (0:00:00.322)       0:03:11.507 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nFriday 13 March 2026  00:50:46 +0000 (0:00:00.291)       0:03:11.799 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nFriday 13 March 2026  00:50:46 +0000 (0:00:00.214)       0:03:12.013 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nFriday 13 March 2026  00:50:46 +0000 (0:00:00.315)       0:03:12.329 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nFriday 13 March 2026  00:50:46 +0000 (0:00:00.180)       0:03:12.510 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nFriday 13 March 2026  00:50:47 +0000 (0:00:00.175)       0:03:12.685 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nFriday 13 March 2026  00:50:47 +0000 (0:00:00.100)       0:03:12.786 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nFriday 13 March 2026  00:50:47 +0000 (0:00:00.585)       0:03:13.371 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nFriday 13 March 2026  00:50:48 +0000 (0:00:00.244)       0:03:13.616 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nFriday 13 March 2026  00:50:48 +0000 (0:00:00.363)       0:03:13.979 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nFriday 13 March 2026  00:50:48 +0000 (0:00:00.278)       0:03:14.258 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nFriday 13 March 2026  00:50:48 +0000 (0:00:00.184)       0:03:14.443 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nFriday 13 March 2026  00:50:49 +0000 (0:00:00.229)       0:03:14.673 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nFriday 13 March 2026  00:50:50 +0000 (0:00:00.854)       0:03:15.528 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nFriday 13 March 2026  00:50:50 +0000 (0:00:00.278)       0:03:15.806 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nFriday 13 March 2026  00:50:51 +0000 (0:00:01.206)       0:03:17.012 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nFriday 13 March 2026  00:50:52 +0000 (0:00:00.841)       0:03:17.854 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nFriday 13 March 2026  00:50:52 +0000 (0:00:00.589)       0:03:18.444 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nFriday 13 March 2026  00:50:53 +0000 (0:00:00.870)       0:03:19.314 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nFriday 13 March 2026  00:50:54 +0000 (0:00:00.532)       0:03:19.847 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nFriday 13 March 2026  00:50:55 +0000 (0:00:01.149)       0:03:20.997 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nFriday 13 March 2026  00:50:56 +0000 (0:00:00.598)       0:03:21.595 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nFriday 13 March 2026  00:50:56 +0000 (0:00:00.327)       0:03:21.923 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nFriday 13 March 2026  00:50:57 +0000 (0:00:00.880)       0:03:22.803 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nFriday 13 March 2026  00:50:58 +0000 (0:00:00.828)       0:03:23.632 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nFriday 13 March 2026  00:50:58 +0000 (0:00:00.562)       0:03:24.194 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nFriday 13 March 2026  00:50:59 +0000 (0:00:01.131)       0:03:25.326 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nFriday 13 March 2026  00:51:00 +0000 (0:00:00.636)       0:03:25.963 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mon] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:51:01 +0000 (0:00:00.601)       0:03:26.564 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:51:01 +0000 (0:00:00.781)       0:03:27.346 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:51:02 +0000 (0:00:00.586)       0:03:27.932 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:51:03 +0000 (0:00:01.090)       0:03:29.023 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:51:03 +0000 (0:00:00.300)       0:03:29.323 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:51:04 +0000 (0:00:00.296)       0:03:29.620 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:51:04 +0000 (0:00:00.239)       0:03:29.860 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:51:05 +0000 (0:00:00.821)       0:03:30.681 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:51:05 +0000 (0:00:00.256)       0:03:30.938 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:51:05 +0000 (0:00:00.304)       0:03:31.243 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:51:06 +0000 (0:00:00.730)       0:03:31.974 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:51:07 +0000 (0:00:00.769)       0:03:32.743 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:51:07 +0000 (0:00:00.267)       0:03:33.010 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:51:07 +0000 (0:00:00.287)       0:03:33.298 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:51:08 +0000 (0:00:00.283)       0:03:33.581 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:51:08 +0000 (0:00:00.250)       0:03:33.832 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:51:08 +0000 (0:00:00.439)       0:03:34.272 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:51:09 +0000 (0:00:00.266)       0:03:34.539 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:51:09 +0000 (0:00:00.254)       0:03:34.793 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:51:09 +0000 (0:00:00.264)       0:03:35.058 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:51:09 +0000 (0:00:00.415)       0:03:35.473 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nFriday 13 March 2026  00:51:10 +0000 (0:00:00.499)       0:03:35.973 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nFriday 13 March 2026  00:51:10 +0000 (0:00:00.284)       0:03:36.258 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nFriday 13 March 2026  00:51:11 +0000 (0:00:00.708)       0:03:36.967 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nFriday 13 March 2026  00:51:11 +0000 (0:00:00.147)       0:03:37.114 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nFriday 13 March 2026  00:51:12 +0000 (0:00:01.096)       0:03:38.210 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nFriday 13 March 2026  00:51:13 +0000 (0:00:00.336)       0:03:38.547 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nFriday 13 March 2026  00:51:13 +0000 (0:00:00.545)       0:03:39.092 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nFriday 13 March 2026  00:51:14 +0000 (0:00:01.163)       0:03:40.256 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nFriday 13 March 2026  00:51:15 +0000 (0:00:00.790)       0:03:41.046 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nFriday 13 March 2026  00:51:16 +0000 (0:00:00.695)       0:03:41.742 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nFriday 13 March 2026  00:51:16 +0000 (0:00:00.635)       0:03:42.377 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nFriday 13 March 2026  00:51:18 +0000 (0:00:01.575)       0:03:43.952 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nFriday 13 March 2026  00:51:19 +0000 (0:00:00.706)       0:03:44.659 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nFriday 13 March 2026  00:51:22 +0000 (0:00:03.007)       0:03:47.666 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nFriday 13 March 2026  00:51:23 +0000 (0:00:01.304)       0:03:48.971 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nFriday 13 March 2026  00:51:23 +0000 (0:00:00.339)       0:03:49.310 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nFriday 13 March 2026  00:51:24 +0000 (0:00:00.507)       0:03:49.818 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nFriday 13 March 2026  00:51:25 +0000 (0:00:01.534)       0:03:51.352 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nFriday 13 March 2026  00:51:27 +0000 (0:00:01.209)       0:03:52.561 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nFriday 13 March 2026  00:51:27 +0000 (0:00:00.218)       0:03:52.780 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nFriday 13 March 2026  00:51:27 +0000 (0:00:00.594)       0:03:53.374 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nFriday 13 March 2026  00:51:28 +0000 (0:00:00.284)       0:03:53.659 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nFriday 13 March 2026  00:51:28 +0000 (0:00:00.255)       0:03:53.914 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nFriday 13 March 2026  00:51:29 +0000 (0:00:00.638)       0:03:54.553 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nFriday 13 March 2026  00:51:31 +0000 (0:00:01.980)       0:03:56.534 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nFriday 13 March 2026  00:51:32 +0000 (0:00:01.205)       0:03:57.739 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nFriday 13 March 2026  00:51:33 +0000 (0:00:01.737)       0:03:59.477 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nFriday 13 March 2026  00:51:36 +0000 (0:00:02.648)       0:04:02.125 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nFriday 13 March 2026  00:51:37 +0000 (0:00:00.542)       0:04:02.668 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nFriday 13 March 2026  00:51:58 +0000 (0:00:21.791)       0:04:24.459 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nFriday 13 March 2026  00:52:07 +0000 (0:00:08.523)       0:04:32.983 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nFriday 13 March 2026  00:52:08 +0000 (0:00:00.567)       0:04:33.551 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}]) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:52:22 +0000 (0:00:14.641)       0:04:48.193 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nFriday 13 March 2026  00:52:22 +0000 (0:00:00.314)       0:04:48.508 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nFriday 13 March 2026  00:52:23 +0000 (0:00:00.816)       0:04:49.324 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nFriday 13 March 2026  00:52:24 +0000 (0:00:00.316)       0:04:49.640 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nFriday 13 March 2026  00:52:24 +0000 (0:00:00.330)       0:04:49.971 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nFriday 13 March 2026  00:52:25 +0000 (0:00:00.870)       0:04:50.842 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mgr] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:52:26 +0000 (0:00:00.793)       0:04:51.636 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:52:26 +0000 (0:00:00.488)       0:04:52.124 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:52:27 +0000 (0:00:00.750)       0:04:52.874 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:52:28 +0000 (0:00:00.754)       0:04:53.629 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:52:28 +0000 (0:00:00.270)       0:04:53.899 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:52:28 +0000 (0:00:00.426)       0:04:54.325 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:52:29 +0000 (0:00:00.338)       0:04:54.664 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:52:29 +0000 (0:00:00.682)       0:04:55.346 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:52:30 +0000 (0:00:00.263)       0:04:55.609 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:52:30 +0000 (0:00:00.420)       0:04:56.029 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:52:31 +0000 (0:00:00.705)       0:04:56.734 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:52:31 +0000 (0:00:00.648)       0:04:57.383 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:52:32 +0000 (0:00:00.234)       0:04:57.617 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:52:32 +0000 (0:00:00.475)       0:04:58.093 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:52:32 +0000 (0:00:00.243)       0:04:58.337 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:52:33 +0000 (0:00:00.270)       0:04:58.607 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:52:33 +0000 (0:00:00.357)       0:04:58.965 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:52:33 +0000 (0:00:00.286)       0:04:59.252 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:52:34 +0000 (0:00:00.550)       0:04:59.802 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:52:34 +0000 (0:00:00.359)       0:05:00.161 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:52:34 +0000 (0:00:00.289)       0:05:00.451 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nFriday 13 March 2026  00:52:35 +0000 (0:00:00.589)       0:05:01.040 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.542)       0:05:01.583 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nFriday 13 March 2026  00:52:36 +0000 (0:00:00.430)       0:05:02.013 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nFriday 13 March 2026  00:52:37 +0000 (0:00:00.594)       0:05:02.608 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nFriday 13 March 2026  00:52:37 +0000 (0:00:00.408)       0:05:03.016 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nFriday 13 March 2026  00:52:47 +0000 (0:00:09.626)       0:05:12.642 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nFriday 13 March 2026  00:52:47 +0000 (0:00:00.285)       0:05:12.928 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nFriday 13 March 2026  00:52:49 +0000 (0:00:01.914)       0:05:14.843 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nFriday 13 March 2026  00:52:50 +0000 (0:00:01.502)       0:05:16.346 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nFriday 13 March 2026  00:52:51 +0000 (0:00:00.779)       0:05:17.125 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nFriday 13 March 2026  00:52:51 +0000 (0:00:00.291)       0:05:17.417 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nFriday 13 March 2026  00:52:52 +0000 (0:00:00.268)       0:05:17.685 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nFriday 13 March 2026  00:52:52 +0000 (0:00:00.568)       0:05:18.254 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nFriday 13 March 2026  00:52:53 +0000 (0:00:00.322)       0:05:18.577 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nFriday 13 March 2026  00:52:53 +0000 (0:00:00.267)       0:05:18.844 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nFriday 13 March 2026  00:52:53 +0000 (0:00:00.496)       0:05:19.340 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nFriday 13 March 2026  00:52:54 +0000 (0:00:01.064)       0:05:20.405 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nFriday 13 March 2026  00:52:56 +0000 (0:00:01.182)       0:05:21.587 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nFriday 13 March 2026  00:52:58 +0000 (0:00:01.967)       0:05:23.555 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nFriday 13 March 2026  00:53:00 +0000 (0:00:02.572)       0:05:26.127 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nFriday 13 March 2026  00:53:01 +0000 (0:00:00.423)       0:05:26.551 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nFriday 13 March 2026  00:53:37 +0000 (0:00:36.006)       0:06:02.557 ********** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nFriday 13 March 2026  00:53:38 +0000 (0:00:01.368)       0:06:03.926 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nFriday 13 March 2026  00:53:38 +0000 (0:00:00.304)       0:06:04.231 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nFriday 13 March 2026  00:53:38 +0000 (0:00:00.107)       0:06:04.338 ********** \n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nFriday 13 March 2026  00:53:45 +0000 (0:00:06.638)       0:06:10.976 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:53:50 +0000 (0:00:04.679)       0:06:15.656 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nFriday 13 March 2026  00:53:50 +0000 (0:00:00.593)       0:06:16.250 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nFriday 13 March 2026  00:53:51 +0000 (0:00:00.576)       0:06:16.827 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nFriday 13 March 2026  00:53:51 +0000 (0:00:00.266)       0:06:17.093 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nFriday 13 March 2026  00:53:52 +0000 (0:00:00.996)       0:06:18.090 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nFriday 13 March 2026  00:53:53 +0000 (0:00:00.525)       0:06:18.616 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-osd] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:53:53 +0000 (0:00:00.663)       0:06:19.279 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:53:54 +0000 (0:00:00.441)       0:06:19.721 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:53:54 +0000 (0:00:00.592)       0:06:20.313 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:53:55 +0000 (0:00:00.258)       0:06:20.572 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:53:55 +0000 (0:00:00.670)       0:06:21.243 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:53:56 +0000 (0:00:00.670)       0:06:21.913 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:53:57 +0000 (0:00:01.022)       0:06:22.936 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:53:57 +0000 (0:00:00.257)       0:06:23.194 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:53:57 +0000 (0:00:00.320)       0:06:23.514 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:53:58 +0000 (0:00:00.223)       0:06:23.738 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:53:59 +0000 (0:00:00.817)       0:06:24.556 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:53:59 +0000 (0:00:00.666)       0:06:25.223 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:53:59 +0000 (0:00:00.208)       0:06:25.432 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:54:00 +0000 (0:00:00.263)       0:06:25.696 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:54:00 +0000 (0:00:00.437)       0:06:26.133 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:54:00 +0000 (0:00:00.280)       0:06:26.414 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:54:01 +0000 (0:00:00.296)       0:06:26.710 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:54:01 +0000 (0:00:00.249)       0:06:26.959 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:54:01 +0000 (0:00:00.409)       0:06:27.368 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:54:02 +0000 (0:00:00.251)       0:06:27.620 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:54:02 +0000 (0:00:00.276)       0:06:27.896 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nFriday 13 March 2026  00:54:02 +0000 (0:00:00.562)       0:06:28.458 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nFriday 13 March 2026  00:54:03 +0000 (0:00:00.279)       0:06:28.738 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nFriday 13 March 2026  00:54:03 +0000 (0:00:00.548)       0:06:29.286 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nFriday 13 March 2026  00:54:04 +0000 (0:00:00.438)       0:06:29.724 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nFriday 13 March 2026  00:54:04 +0000 (0:00:00.399)       0:06:30.124 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nFriday 13 March 2026  00:54:04 +0000 (0:00:00.263)       0:06:30.387 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nFriday 13 March 2026  00:54:05 +0000 (0:00:00.663)       0:06:31.051 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nFriday 13 March 2026  00:54:05 +0000 (0:00:00.429)       0:06:31.481 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nFriday 13 March 2026  00:54:09 +0000 (0:00:03.744)       0:06:35.226 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nFriday 13 March 2026  00:54:10 +0000 (0:00:00.305)       0:06:35.532 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nFriday 13 March 2026  00:54:10 +0000 (0:00:00.403)       0:06:35.936 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nFriday 13 March 2026  00:54:11 +0000 (0:00:01.248)       0:06:37.184 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nFriday 13 March 2026  00:54:13 +0000 (0:00:02.244)       0:06:39.428 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nFriday 13 March 2026  00:54:15 +0000 (0:00:01.282)       0:06:40.711 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nFriday 13 March 2026  00:54:17 +0000 (0:00:02.626)       0:06:43.337 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Use ceph-volume to create osds] *******************************\nFriday 13 March 2026  00:54:18 +0000 (0:00:00.746)       0:06:44.084 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nFriday 13 March 2026  00:55:04 +0000 (0:00:45.469)       0:07:29.554 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nFriday 13 March 2026  00:55:04 +0000 (0:00:00.378)       0:07:29.932 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-4, testbed-node-5, testbed-node-3\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nFriday 13 March 2026  00:55:05 +0000 (0:00:00.839)       0:07:30.772 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nFriday 13 March 2026  00:55:06 +0000 (0:00:00.802)       0:07:31.574 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nFriday 13 March 2026  00:55:09 +0000 (0:00:03.032)       0:07:34.606 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nFriday 13 March 2026  00:55:09 +0000 (0:00:00.707)       0:07:35.314 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nFriday 13 March 2026  00:55:10 +0000 (0:00:01.188)       0:07:36.503 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nFriday 13 March 2026  00:55:12 +0000 (0:00:01.109)       0:07:37.612 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nFriday 13 March 2026  00:55:14 +0000 (0:00:01.927)       0:07:39.539 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nFriday 13 March 2026  00:55:14 +0000 (0:00:00.580)       0:07:40.120 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nFriday 13 March 2026  00:55:14 +0000 (0:00:00.320)       0:07:40.441 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=3)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nFriday 13 March 2026  00:55:16 +0000 (0:00:01.284)       0:07:41.725 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nFriday 13 March 2026  00:55:18 +0000 (0:00:02.365)       0:07:44.091 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=1)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nFriday 13 March 2026  00:55:22 +0000 (0:00:03.566)       0:07:47.657 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nFriday 13 March 2026  00:55:25 +0000 (0:00:02.996)       0:07:50.654 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nFriday 13 March 2026  00:55:37 +0000 (0:00:12.509)       0:08:03.164 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:55:38 +0000 (0:00:00.994)       0:08:04.159 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nFriday 13 March 2026  00:55:38 +0000 (0:00:00.340)       0:08:04.499 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nFriday 13 March 2026  00:55:39 +0000 (0:00:00.459)       0:08:04.959 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nFriday 13 March 2026  00:55:40 +0000 (0:00:00.656)       0:08:05.615 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nFriday 13 March 2026  00:55:40 +0000 (0:00:00.295)       0:08:05.911 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nFriday 13 March 2026  00:55:40 +0000 (0:00:00.217)       0:08:06.129 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nFriday 13 March 2026  00:55:40 +0000 (0:00:00.262)       0:08:06.391 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nFriday 13 March 2026  00:55:41 +0000 (0:00:00.192)       0:08:06.583 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nFriday 13 March 2026  00:55:41 +0000 (0:00:00.198)       0:08:06.782 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nFriday 13 March 2026  00:55:41 +0000 (0:00:00.111)       0:08:06.894 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nFriday 13 March 2026  00:55:41 +0000 (0:00:00.180)       0:08:07.075 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nFriday 13 March 2026  00:55:42 +0000 (0:00:00.561)       0:08:07.637 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nFriday 13 March 2026  00:55:42 +0000 (0:00:00.341)       0:08:07.978 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nFriday 13 March 2026  00:55:42 +0000 (0:00:00.311)       0:08:08.289 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nFriday 13 March 2026  00:55:42 +0000 (0:00:00.183)       0:08:08.473 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role ceph-crash] ***************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:55:43 +0000 (0:00:00.720)       0:08:09.194 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:55:44 +0000 (0:00:00.941)       0:08:10.135 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:55:45 +0000 (0:00:00.877)       0:08:11.013 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:55:46 +0000 (0:00:01.054)       0:08:12.067 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:55:47 +0000 (0:00:00.668)       0:08:12.736 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:55:48 +0000 (0:00:00.802)       0:08:13.539 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:55:48 +0000 (0:00:00.635)       0:08:14.174 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:55:49 +0000 (0:00:01.003)       0:08:15.178 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:55:50 +0000 (0:00:00.510)       0:08:15.688 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:55:50 +0000 (0:00:00.690)       0:08:16.379 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:55:51 +0000 (0:00:00.991)       0:08:17.371 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:55:53 +0000 (0:00:01.383)       0:08:18.755 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:55:53 +0000 (0:00:00.615)       0:08:19.370 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:55:54 +0000 (0:00:00.860)       0:08:20.231 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:55:55 +0000 (0:00:00.570)       0:08:20.801 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:55:56 +0000 (0:00:00.818)       0:08:21.620 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:55:56 +0000 (0:00:00.579)       0:08:22.200 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:55:57 +0000 (0:00:00.777)       0:08:22.977 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:55:58 +0000 (0:00:00.569)       0:08:23.547 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:55:58 +0000 (0:00:00.805)       0:08:24.353 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:55:59 +0000 (0:00:00.596)       0:08:24.949 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nFriday 13 March 2026  00:56:00 +0000 (0:00:01.222)       0:08:26.171 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nFriday 13 March 2026  00:56:04 +0000 (0:00:03.585)       0:08:29.756 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nFriday 13 March 2026  00:56:06 +0000 (0:00:02.244)       0:08:32.001 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nFriday 13 March 2026  00:56:08 +0000 (0:00:01.887)       0:08:33.888 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nFriday 13 March 2026  00:56:09 +0000 (0:00:00.940)       0:08:34.829 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nFriday 13 March 2026  00:56:10 +0000 (0:00:01.245)       0:08:36.074 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nFriday 13 March 2026  00:56:12 +0000 (0:00:01.633)       0:08:37.708 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************\nFriday 13 March 2026  00:56:15 +0000 (0:00:03.046)       0:08:40.754 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******\nFriday 13 March 2026  00:56:16 +0000 (0:00:01.299)       0:08:42.053 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************\nFriday 13 March 2026  00:56:17 +0000 (0:00:00.824)       0:08:42.878 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******\nFriday 13 March 2026  00:56:19 +0000 (0:00:02.209)       0:08:45.087 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mds] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.920)       0:08:46.008 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.454)       0:08:46.462 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:56:21 +0000 (0:00:00.739)       0:08:47.202 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:56:21 +0000 (0:00:00.254)       0:08:47.456 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:56:22 +0000 (0:00:00.634)       0:08:48.091 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:56:23 +0000 (0:00:00.806)       0:08:48.897 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:00.741)       0:08:49.639 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:00.306)       0:08:49.945 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:00.291)       0:08:50.236 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:56:25 +0000 (0:00:00.585)       0:08:50.822 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:56:26 +0000 (0:00:00.806)       0:08:51.629 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:56:26 +0000 (0:00:00.775)       0:08:52.404 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:56:27 +0000 (0:00:00.314)       0:08:52.719 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:56:27 +0000 (0:00:00.575)       0:08:53.294 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:56:28 +0000 (0:00:00.326)       0:08:53.622 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:56:28 +0000 (0:00:00.361)       0:08:53.983 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:56:28 +0000 (0:00:00.343)       0:08:54.326 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:56:29 +0000 (0:00:00.623)       0:08:54.950 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:56:29 +0000 (0:00:00.308)       0:08:55.259 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:56:30 +0000 (0:00:00.309)       0:08:55.568 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:56:30 +0000 (0:00:00.321)       0:08:55.890 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nFriday 13 March 2026  00:56:31 +0000 (0:00:00.776)       0:08:56.666 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule details] *********************\nFriday 13 March 2026  00:56:31 +0000 (0:00:00.426)       0:08:57.093 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule name] ************************\nFriday 13 March 2026  00:56:33 +0000 (0:00:01.888)       0:08:58.981 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create filesystem pools] **************************************\nFriday 13 March 2026  00:56:33 +0000 (0:00:00.242)       0:08:59.224 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [ceph-mds : Create ceph filesystem] ***************************************\nFriday 13 March 2026  00:56:42 +0000 (0:00:08.714)       0:09:07.938 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nFriday 13 March 2026  00:56:46 +0000 (0:00:03.786)       0:09:11.725 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nFriday 13 March 2026  00:56:46 +0000 (0:00:00.451)       0:09:12.177 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nFriday 13 March 2026  00:56:47 +0000 (0:00:01.153)       0:09:13.331 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nFriday 13 March 2026  00:56:50 +0000 (0:00:02.738)       0:09:16.069 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nFriday 13 March 2026  00:56:51 +0000 (0:00:01.385)       0:09:17.455 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nFriday 13 March 2026  00:56:54 +0000 (0:00:02.308)       0:09:19.764 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nFriday 13 March 2026  00:56:54 +0000 (0:00:00.237)       0:09:20.002 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nFriday 13 March 2026  00:56:55 +0000 (0:00:00.620)       0:09:20.623 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nFriday 13 March 2026  00:56:55 +0000 (0:00:00.465)       0:09:21.088 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nFriday 13 March 2026  00:56:56 +0000 (0:00:01.065)       0:09:22.154 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nFriday 13 March 2026  00:56:58 +0000 (0:00:01.412)       0:09:23.566 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nFriday 13 March 2026  00:56:59 +0000 (0:00:01.696)       0:09:25.263 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nFriday 13 March 2026  00:57:01 +0000 (0:00:01.877)       0:09:27.141 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:57:02 +0000 (0:00:01.266)       0:09:28.407 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nFriday 13 March 2026  00:57:03 +0000 (0:00:00.612)       0:09:29.020 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nFriday 13 March 2026  00:57:04 +0000 (0:00:00.697)       0:09:29.718 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nFriday 13 March 2026  00:57:04 +0000 (0:00:00.301)       0:09:30.019 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nFriday 13 March 2026  00:57:05 +0000 (0:00:01.200)       0:09:31.220 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nFriday 13 March 2026  00:57:06 +0000 (0:00:00.658)       0:09:31.879 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nFriday 13 March 2026  00:57:06 +0000 (0:00:00.575)       0:09:32.455 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nFriday 13 March 2026  00:57:07 +0000 (0:00:00.437)       0:09:32.892 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nFriday 13 March 2026  00:57:07 +0000 (0:00:00.538)       0:09:33.431 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nFriday 13 March 2026  00:57:08 +0000 (0:00:00.257)       0:09:33.688 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nFriday 13 March 2026  00:57:09 +0000 (0:00:00.869)       0:09:34.557 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nFriday 13 March 2026  00:57:09 +0000 (0:00:00.949)       0:09:35.507 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nFriday 13 March 2026  00:57:10 +0000 (0:00:00.697)       0:09:36.204 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nFriday 13 March 2026  00:57:10 +0000 (0:00:00.259)       0:09:36.463 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nFriday 13 March 2026  00:57:11 +0000 (0:00:00.288)       0:09:36.752 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nFriday 13 March 2026  00:57:11 +0000 (0:00:00.413)       0:09:37.166 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nFriday 13 March 2026  00:57:12 +0000 (0:00:00.650)       0:09:37.817 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nFriday 13 March 2026  00:57:12 +0000 (0:00:00.605)       0:09:38.422 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nFriday 13 March 2026  00:57:13 +0000 (0:00:00.256)       0:09:38.679 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nFriday 13 March 2026  00:57:13 +0000 (0:00:00.409)       0:09:39.088 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nFriday 13 March 2026  00:57:13 +0000 (0:00:00.274)       0:09:39.362 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nFriday 13 March 2026  00:57:14 +0000 (0:00:00.276)       0:09:39.639 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nFriday 13 March 2026  00:57:14 +0000 (0:00:00.279)       0:09:39.919 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nFriday 13 March 2026  00:57:14 +0000 (0:00:00.414)       0:09:40.333 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nFriday 13 March 2026  00:57:15 +0000 (0:00:00.276)       0:09:40.609 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nFriday 13 March 2026  00:57:15 +0000 (0:00:00.295)       0:09:40.905 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nFriday 13 March 2026  00:57:15 +0000 (0:00:00.280)       0:09:41.185 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nFriday 13 March 2026  00:57:16 +0000 (0:00:00.587)       0:09:41.772 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nFriday 13 March 2026  00:57:16 +0000 (0:00:00.456)       0:09:42.229 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nFriday 13 March 2026  00:57:18 +0000 (0:00:01.973)       0:09:44.202 ********** \n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nFriday 13 March 2026  00:57:20 +0000 (0:00:01.547)       0:09:45.749 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nFriday 13 March 2026  00:57:20 +0000 (0:00:00.315)       0:09:46.065 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nFriday 13 March 2026  00:57:21 +0000 (0:00:00.515)       0:09:46.580 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nFriday 13 March 2026  00:57:22 +0000 (0:00:01.426)       0:09:48.007 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nFriday 13 March 2026  00:57:26 +0000 (0:00:04.077)       0:09:52.084 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nFriday 13 March 2026  00:57:28 +0000 (0:00:02.092)       0:09:54.177 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nFriday 13 March 2026  00:57:30 +0000 (0:00:01.382)       0:09:55.559 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nFriday 13 March 2026  00:57:30 +0000 (0:00:00.223)       0:09:55.782 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nFriday 13 March 2026  00:57:31 +0000 (0:00:01.154)       0:09:56.937 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nFriday 13 March 2026  00:57:32 +0000 (0:00:00.598)       0:09:57.535 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nFriday 13 March 2026  00:58:02 +0000 (0:00:30.025)       0:10:27.561 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nFriday 13 March 2026  00:58:02 +0000 (0:00:00.254)       0:10:27.816 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nFriday 13 March 2026  00:58:02 +0000 (0:00:00.298)       0:10:28.114 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nFriday 13 March 2026  00:58:03 +0000 (0:00:00.611)       0:10:28.725 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nFriday 13 March 2026  00:58:03 +0000 (0:00:00.460)       0:10:29.186 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nFriday 13 March 2026  00:58:04 +0000 (0:00:01.133)       0:10:30.319 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nFriday 13 March 2026  00:58:06 +0000 (0:00:01.226)       0:10:31.546 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nFriday 13 March 2026  00:58:07 +0000 (0:00:01.607)       0:10:33.154 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nFriday 13 March 2026  00:58:10 +0000 (0:00:03.299)       0:10:36.454 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nFriday 13 March 2026  00:58:11 +0000 (0:00:00.303)       0:10:36.758 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nFriday 13 March 2026  00:58:11 +0000 (0:00:00.444)       0:10:37.202 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nFriday 13 March 2026  00:58:12 +0000 (0:00:00.446)       0:10:37.649 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nFriday 13 March 2026  00:58:12 +0000 (0:00:00.296)       0:10:37.945 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nFriday 13 March 2026  00:58:12 +0000 (0:00:00.533)       0:10:38.479 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=32  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:58:13 +0000 (0:00:00.207)       0:10:38.687 ********** \n=============================================================================== \nceph-osd : Use ceph-volume to create osds ------------------------------ 45.47s\nceph-container-common : Pulling Ceph container image ------------------- 42.05s\nceph-mgr : Wait for all mgr to be up ----------------------------------- 36.01s\nceph-rgw : Create rgw pools -------------------------------------------- 30.03s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.79s\nceph-mon : Set cluster configs ----------------------------------------- 14.64s\nceph-osd : Wait for all osd to be up ----------------------------------- 12.51s\nceph-mgr : Create ceph mgr keyring(s) on a mon node --------------------- 9.63s\nceph-mds : Create filesystem pools -------------------------------------- 8.71s\nceph-mon : Fetch ceph initial keys -------------------------------------- 8.52s\nceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.64s\nceph-config : Create ceph initial directories --------------------------- 6.46s\nceph-mgr : Add modules to ceph-mgr -------------------------------------- 4.68s\nceph-rgw : Create rgw keyrings ------------------------------------------ 4.08s\nceph-mds : Create ceph filesystem --------------------------------------- 3.79s\nceph-osd : Apply operating system tuning -------------------------------- 3.75s\nceph-crash : Create client.crash keyring -------------------------------- 3.59s\nceph-osd : Systemd start osd -------------------------------------------- 3.57s\nceph-container-common : Get ceph version -------------------------------- 3.52s\nceph-container-common : Enable ceph.target ------------------------------ 3.48s\n\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.248)       0:00:00.248 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.317)       0:00:00.565 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nFriday 13 March 2026  00:56:21 +0000 (0:00:00.439)       0:00:01.005 ********** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nFriday 13 March 2026  00:56:21 +0000 (0:00:00.467)       0:00:01.472 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:02.577)       0:00:04.049 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nFriday 13 March 2026  00:56:26 +0000 (0:00:02.173)       0:00:06.223 ********** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nFriday 13 March 2026  00:56:27 +0000 (0:00:00.946)       0:00:07.170 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:56:30 +0000 (0:00:03.011)       0:00:10.182 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nFriday 13 March 2026  00:56:32 +0000 (0:00:01.529)       0:00:11.712 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nFriday 13 March 2026  00:56:33 +0000 (0:00:01.285)       0:00:12.997 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nFriday 13 March 2026  00:56:36 +0000 (0:00:02.631)       0:00:15.629 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nFriday 13 March 2026  00:56:38 +0000 (0:00:02.498)       0:00:18.127 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : opensearch | Check containers] ****************\nFriday 13 March 2026  00:56:40 +0000 (0:00:02.125)       0:00:20.253 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m\n\nTASK [service-check-containers : opensearch | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:56:42 +0000 (0:00:01.844)       0:00:22.098 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:56:42 +0000 (0:00:00.450)       0:00:22.548 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:01.233)       0:00:23.782 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:00.303)       0:00:24.085 ********** \n\nTASK [opensearch : Flush handlers] *********************************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:00.063)       0:00:24.148 ********** \n\nTASK [opensearch : Flush handlers] *********************************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:00.063)       0:00:24.212 ********** \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:00.065)       0:00:24.277 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nFriday 13 March 2026  00:56:44 +0000 (0:00:00.195)       0:00:24.472 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nFriday 13 March 2026  00:56:45 +0000 (0:00:00.217)       0:00:24.690 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nFriday 13 March 2026  00:57:37 +0000 (0:00:52.408)       0:01:17.098 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nFriday 13 March 2026  00:58:35 +0000 (0:00:57.632)       0:02:14.730 ********** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nFriday 13 March 2026  00:58:35 +0000 (0:00:00.543)       0:02:15.273 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch cluster to become healthy] **************\nFriday 13 March 2026  00:58:37 +0000 (0:00:02.169)       0:02:17.443 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nFriday 13 March 2026  00:58:39 +0000 (0:00:01.854)       0:02:19.297 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nFriday 13 March 2026  00:58:42 +0000 (0:00:02.462)       0:02:21.760 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nFriday 13 March 2026  00:58:44 +0000 (0:00:02.214)       0:02:23.974 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:58:46 +0000 (0:00:02.050)       0:02:26.024 ********** \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 57.63s\nopensearch : Restart opensearch container ------------------------------ 52.41s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 3.01s\nopensearch : Copying over config.json files for services ---------------- 2.63s\nopensearch : Setting sysctl values -------------------------------------- 2.58s\nopensearch : Copying over opensearch service config file ---------------- 2.50s\nopensearch : Check if a log retention policy exists --------------------- 2.46s\nopensearch : Create new log retention policy ---------------------------- 2.21s\nopensearch : Ensuring config directories exist -------------------------- 2.17s\nopensearch : Wait for OpenSearch to become ready ------------------------ 2.17s\nopensearch : Copying over opensearch-dashboards config file ------------- 2.13s\nopensearch : Apply retention policy to existing indices ----------------- 2.05s\nopensearch : Wait for OpenSearch cluster to become healthy -------------- 1.85s\nservice-check-containers : opensearch | Check containers ---------------- 1.84s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.53s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 1.29s\nservice-check-containers : Include tasks -------------------------------- 1.23s\nopensearch : include_tasks ---------------------------------------------- 0.95s\nopensearch : include_tasks ---------------------------------------------- 0.54s\nopensearch : include_tasks ---------------------------------------------- 0.47s\n\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state SUCCESS\u001b[0m\n\n\nPLAY [Set kolla_action_mariadb] ************************************************\n\nTASK [Inform the user about the following task] ********************************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.102)       0:00:00.102 ********** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check MariaDB service] ***************************************************\nFriday 13 March 2026  00:56:20 +0000 (0:00:00.030)       0:00:00.133 ********** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********\nFriday 13 March 2026  00:56:23 +0000 (0:00:02.861)       0:00:02.994 ********** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_mariadb = kolla_action_ng] ******************************\nFriday 13 March 2026  00:56:23 +0000 (0:00:00.053)       0:00:03.048 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:56:23 +0000 (0:00:00.158)       0:00:03.207 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:00.324)       0:00:03.532 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nFriday 13 March 2026  00:56:24 +0000 (0:00:00.606)       0:00:04.138 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nFriday 13 March 2026  00:56:25 +0000 (0:00:00.390)       0:00:04.528 ********** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nFriday 13 March 2026  00:56:25 +0000 (0:00:00.618)       0:00:05.147 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nFriday 13 March 2026  00:56:29 +0000 (0:00:03.402)       0:00:08.549 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nFriday 13 March 2026  00:56:29 +0000 (0:00:00.640)       0:00:09.190 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nFriday 13 March 2026  00:56:31 +0000 (0:00:01.488)       0:00:10.678 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nFriday 13 March 2026  00:56:35 +0000 (0:00:03.842)       0:00:14.521 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nFriday 13 March 2026  00:56:36 +0000 (0:00:01.124)       0:00:15.646 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nFriday 13 March 2026  00:56:40 +0000 (0:00:04.183)       0:00:19.829 ********** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nFriday 13 March 2026  00:56:40 +0000 (0:00:00.544)       0:00:20.373 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:56:43 +0000 (0:00:02.661)       0:00:23.035 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nFriday 13 March 2026  00:56:46 +0000 (0:00:02.524)       0:00:25.559 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : mariadb | Check containers] *******************\nFriday 13 March 2026  00:56:48 +0000 (0:00:02.688)       0:00:28.248 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/l\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\nocaltime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [service-check-containers : mariadb | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:56:51 +0000 (0:00:02.508)       0:00:30.756 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:56:51 +0000 (0:00:00.561)       0:00:31.318 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Checking for mariadb cluster] **********************************\nFriday 13 March 2026  00:56:54 +0000 (0:00:02.354)       0:00:33.672 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Cleaning up temp file on localhost] ****************************\nFriday 13 March 2026  00:56:54 +0000 (0:00:00.331)       0:00:34.004 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Stop MariaDB containers] ***************************************\nFriday 13 March 2026  00:56:54 +0000 (0:00:00.159)       0:00:34.163 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Run MariaDB wsrep recovery] ************************************\nFriday 13 March 2026  00:56:55 +0000 (0:00:00.490)       0:00:34.653 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Copying MariaDB log file to /tmp] ******************************\nFriday 13 March 2026  00:56:55 +0000 (0:00:00.365)       0:00:35.019 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Get MariaDB wsrep recovery seqno] ******************************\nFriday 13 March 2026  00:56:55 +0000 (0:00:00.329)       0:00:35.348 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Removing MariaDB log file from /tmp] ***************************\nFriday 13 March 2026  00:56:56 +0000 (0:00:00.284)       0:00:35.633 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Registering MariaDB seqno variable] ****************************\nFriday 13 March 2026  00:56:56 +0000 (0:00:00.501)       0:00:36.134 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Comparing seqno value on all mariadb hosts] ********************\nFriday 13 March 2026  00:56:56 +0000 (0:00:00.320)       0:00:36.454 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Writing hostname of host with the largest seqno to temp file] ***\nFriday 13 March 2026  00:56:57 +0000 (0:00:00.321)       0:00:36.775 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Registering mariadb_recover_inventory_name from temp file] *****\nFriday 13 March 2026  00:56:57 +0000 (0:00:00.288)       0:00:37.064 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Store bootstrap and master hostnames into facts] ***************\nFriday 13 March 2026  00:56:57 +0000 (0:00:00.400)       0:00:37.464 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Set grastate.dat file from MariaDB container in bootstrap host] ***\nFriday 13 March 2026  00:56:58 +0000 (0:00:00.288)       0:00:37.753 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Starting first MariaDB container] ******************************\nFriday 13 March 2026  00:56:58 +0000 (0:00:00.288)       0:00:38.041 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for first MariaDB container] ******************************\nFriday 13 March 2026  00:56:58 +0000 (0:00:00.284)       0:00:38.326 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Set first MariaDB container as primary] ************************\nFriday 13 March 2026  00:56:59 +0000 (0:00:00.306)       0:00:38.632 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB to become operational] ************************\nFriday 13 March 2026  00:56:59 +0000 (0:00:00.430)       0:00:39.063 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Restart slave MariaDB container(s)] ****************************\nFriday 13 March 2026  00:56:59 +0000 (0:00:00.278)       0:00:39.342 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for slave MariaDB] ****************************************\nFriday 13 March 2026  00:57:02 +0000 (0:00:02.296)       0:00:41.638 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Restart master MariaDB container(s)] ***************************\nFriday 13 March 2026  00:57:02 +0000 (0:00:00.283)       0:00:41.922 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for master mariadb] ***************************************\nFriday 13 March 2026  00:57:04 +0000 (0:00:02.378)       0:00:44.301 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Get container facts] ***************************\nFriday 13 March 2026  00:57:05 +0000 (0:00:00.281)       0:00:44.582 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are missing or not running] ***\nFriday 13 March 2026  00:57:05 +0000 (0:00:00.291)       0:00:44.873 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are unhealthy] **************\nFriday 13 March 2026  00:57:05 +0000 (0:00:00.279)       0:00:45.153 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nFriday 13 March 2026  00:57:06 +0000 (0:00:00.579)       0:00:45.733 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nFriday 13 March 2026  00:57:06 +0000 (0:00:00.284)       0:00:46.018 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nFriday 13 March 2026  00:57:07 +0000 (0:00:00.952)       0:00:46.970 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nFriday 13 March 2026  00:57:07 +0000 (0:00:00.398)       0:00:47.369 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nFriday 13 March 2026  00:57:08 +0000 (0:00:00.300)       0:00:47.670 ********** \n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nFriday 13 March 2026  00:57:18 +0000 (0:00:10.775)       0:00:58.445 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nFriday 13 March 2026  00:57:19 +0000 (0:00:00.286)       0:00:58.731 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nFriday 13 March 2026  00:57:19 +0000 (0:00:00.404)       0:00:59.136 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nFriday 13 March 2026  00:57:19 +0000 (0:00:00.325)       0:00:59.462 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nFriday 13 March 2026  00:57:20 +0000 (0:00:00.321)       0:00:59.784 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nFriday 13 March 2026  00:57:20 +0000 (0:00:00.326)       0:01:00.110 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nFriday 13 March 2026  00:57:21 +0000 (0:00:00.496)       0:01:00.606 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [mariadb : Running MariaDB bootstrap container] ***************************\nFriday 13 March 2026  00:57:21 +0000 (0:00:00.366)       0:01:00.972 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Store bootstrap host name into facts] **************************\nFriday 13 March 2026  00:57:31 +0000 (0:00:09.987)       0:01:10.959 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nFriday 13 March 2026  00:57:31 +0000 (0:00:00.116)       0:01:11.076 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Starting first MariaDB container] *******************\nFriday 13 March 2026  00:57:32 +0000 (0:00:00.850)       0:01:11.926 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******\nFriday 13 March 2026  00:57:39 +0000 (0:00:07.389)       0:01:19.315 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******\nFriday 13 March 2026  00:57:41 +0000 (0:00:01.726)       0:01:21.042 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***\nFriday 13 March 2026  00:57:43 +0000 (0:00:02.167)       0:01:23.210 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nFriday 13 March 2026  00:57:43 +0000 (0:00:00.132)       0:01:23.342 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nFriday 13 March 2026  00:57:44 +0000 (0:00:00.363)       0:01:23.706 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nFriday 13 March 2026  00:57:44 +0000 (0:00:00.551)       0:01:24.257 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nFriday 13 March 2026  00:57:58 +0000 (0:00:14.076)       0:01:38.334 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nFriday 13 March 2026  00:58:13 +0000 (0:00:14.503)       0:01:52.838 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nFriday 13 March 2026  00:58:15 +0000 (0:00:01.887)       0:01:54.726 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nFriday 13 March 2026  00:58:35 +0000 (0:00:19.917)       0:02:14.643 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nFriday 13 March 2026  00:58:44 +0000 (0:00:09.573)       0:02:24.217 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nFriday 13 March 2026  00:58:46 +0000 (0:00:01.868)       0:02:26.085 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nFriday 13 March 2026  00:58:56 +0000 (0:00:09.776)       0:02:35.862 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nFriday 13 March 2026  00:59:00 +0000 (0:00:04.597)       0:02:40.460 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nFriday 13 March 2026  00:59:03 +0000 (0:00:02.110)       0:02:42.571 ********** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nFriday 13 March 2026  00:59:03 +0000 (0:00:00.473)       0:02:43.045 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nFriday 13 March 2026  00:59:06 +0000 (0:00:02.468)       0:02:45.513 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nFriday 13 March 2026  00:59:08 +0000 (0:00:02.635)       0:02:48.148 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nFriday 13 March 2026  00:59:10 +0000 (0:00:02.132)       0:02:50.281 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-check : mariadb | Get container facts] ***************************\nFriday 13 March 2026  00:59:12 +0000 (0:00:02.011)       0:02:52.292 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are missing or not running] ***\nFriday 13 March 2026  00:59:17 +0000 (0:00:04.405)       0:02:56.698 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check : mariadb | Fail if containers are unhealthy] **************\nFriday 13 March 2026  00:59:19 +0000 (0:00:02.355)       0:02:59.054 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nFriday 13 March 2026  00:59:20 +0000 (0:00:00.523)       0:02:59.577 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nFriday 13 March 2026  00:59:22 +0000 (0:00:02.659)       0:03:02.237 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=36  \u001b[0m \u001b[0;33mchanged=17  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=39  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  00:59:23 +0000 (0:00:00.400)       0:03:02.637 ********** \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 34.00s\nmariadb : Wait for MariaDB service port liveness ----------------------- 24.08s\nmariadb : Check MariaDB service port liveness -------------------------- 10.78s\nmariadb : Running MariaDB bootstrap container --------------------------- 9.99s\nmariadb : Restart MariaDB container ------------------------------------- 9.78s\nmariadb : Starting first MariaDB container ------------------------------ 7.39s\nmariadb : Wait for MariaDB service port liveness ------------------------ 4.60s\nservice-check : mariadb | Get container facts --------------------------- 4.41s\nmariadb : Copying over galera.cnf --------------------------------------- 4.18s\nmariadb : Copying over config.json files for services ------------------- 3.84s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 3.76s\nmariadb : Ensuring config directories exist ----------------------------- 3.40s\nCheck MariaDB service --------------------------------------------------- 2.86s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.69s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 2.66s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 2.66s\nmariadb : Creating mysql monitor user ----------------------------------- 2.64s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.52s\nservice-check-containers : mariadb | Check containers ------------------- 2.51s\nmariadb : Creating shard root mysql user -------------------------------- 2.47s\n\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Create ceph pools] *******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nFriday 13 March 2026  00:58:17 +0000 (0:00:00.537)       0:00:00.537 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nFriday 13 March 2026  00:58:18 +0000 (0:00:00.532)       0:00:01.070 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nFriday 13 March 2026  00:58:18 +0000 (0:00:00.566)       0:00:01.636 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nFriday 13 March 2026  00:58:19 +0000 (0:00:00.274)       0:00:01.911 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nFriday 13 March 2026  00:58:19 +0000 (0:00:00.687)       0:00:02.599 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nFriday 13 March 2026  00:58:19 +0000 (0:00:00.260)       0:00:02.860 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nFriday 13 March 2026  00:58:20 +0000 (0:00:00.286)       0:00:03.146 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nFriday 13 March 2026  00:58:20 +0000 (0:00:00.307)       0:00:03.454 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nFriday 13 March 2026  00:58:20 +0000 (0:00:00.392)       0:00:03.846 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nFriday 13 March 2026  00:58:21 +0000 (0:00:00.249)       0:00:04.096 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nFriday 13 March 2026  00:58:21 +0000 (0:00:00.583)       0:00:04.680 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nFriday 13 March 2026  00:58:22 +0000 (0:00:00.372)       0:00:05.052 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nFriday 13 March 2026  00:58:23 +0000 (0:00:01.808)       0:00:06.861 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nFriday 13 March 2026  00:58:24 +0000 (0:00:00.462)       0:00:07.323 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nFriday 13 March 2026  00:58:25 +0000 (0:00:00.856)       0:00:08.179 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nFriday 13 March 2026  00:58:25 +0000 (0:00:00.267)       0:00:08.446 ********** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '8f8762892bf0', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-13 00:58:22.722015', 'end': '2026-03-13 00:58:22.745024', 'delta': '0:00:00.023009', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['8f8762892bf0'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'cb0a6fe177a7', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-13 00:58:23.348799', 'end': '2026-03-13 00:58:23.374570', 'delta': '0:00:00.025771', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['cb0a6fe177a7'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '73131559e39b', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-13 00:58:23.809996', 'end': '2026-03-13 00:58:23.845030', 'delta': '0:00:00.035034', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['73131559e39b'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nFriday 13 March 2026  00:58:25 +0000 (0:00:00.177)       0:00:08.623 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nFriday 13 March 2026  00:58:26 +0000 (0:00:00.377)       0:00:09.001 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nFriday 13 March 2026  00:58:27 +0000 (0:00:01.888)       0:00:10.889 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nFriday 13 March 2026  00:58:28 +0000 (0:00:00.280)       0:00:11.170 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nFriday 13 March 2026  00:58:28 +0000 (0:00:00.350)       0:00:11.521 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nFriday 13 March 2026  00:58:29 +0000 (0:00:00.382)       0:00:11.903 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nFriday 13 March 2026  00:58:29 +0000 (0:00:00.108)       0:00:12.012 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nFriday 13 March 2026  00:58:29 +0000 (0:00:00.230)       0:00:12.243 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nFriday 13 March 2026  00:58:29 +0000 (0:00:00.253)       0:00:12.497 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nFriday 13 March 2026  00:58:29 +0000 (0:00:00.279)       0:00:12.776 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nFriday 13 March 2026  00:58:30 +0000 (0:00:00.383)       0:00:13.160 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nFriday 13 March 2026  00:58:30 +0000 (0:00:00.296)       0:00:13.456 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nFriday 13 March 2026  00:58:30 +0000 (0:00:00.262)       0:00:13.719 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nFriday 13 March 2026  00:58:31 +0000 (0:00:00.271)       0:00:13.991 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nFriday 13 March 2026  00:58:31 +0000 (0:00:00.442)       0:00:14.434 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nFriday 13 March 2026  00:58:32 +0000 (0:00:00.555)       0:00:14.989 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nFriday 13 March 2026  00:58:32 +0000 (0:00:00.539)       0:00:15.529 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nFriday 13 March 2026  00:58:33 +0000 (0:00:00.615)       0:00:16.145 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nFriday 13 March 2026  00:58:33 +0000 (0:00:00.352)       0:00:16.497 ********** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nFriday 13 March 2026  00:58:34 +0000 (0:00:00.553)       0:00:17.051 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nFriday 13 March 2026  00:58:34 +0000 (0:00:00.213)       0:00:17.264 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nFriday 13 March 2026  00:58:34 +0000 (0:00:00.325)       0:00:17.590 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nFriday 13 March 2026  00:58:35 +0000 (0:00:00.353)       0:00:17.944 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nFriday 13 March 2026  00:58:35 +0000 (0:00:00.805)       0:00:18.749 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nFriday 13 March 2026  00:58:36 +0000 (0:00:00.301)       0:00:19.051 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nFriday 13 March 2026  00:58:36 +0000 (0:00:00.555)       0:00:19.606 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nFriday 13 March 2026  00:58:36 +0000 (0:00:00.264)       0:00:19.871 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nFriday 13 March 2026  00:58:37 +0000 (0:00:00.274)       0:00:20.145 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nFriday 13 March 2026  00:58:37 +0000 (0:00:00.271)       0:00:20.417 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nFriday 13 March 2026  00:58:38 +0000 (0:00:00.480)       0:00:20.897 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nFriday 13 March 2026  00:58:38 +0000 (0:00:00.361)       0:00:21.259 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nFriday 13 March 2026  00:58:38 +0000 (0:00:00.343)       0:00:21.602 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nFriday 13 March 2026  00:58:39 +0000 (0:00:00.330)       0:00:21.933 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nFriday 13 March 2026  00:58:39 +0000 (0:00:00.264)       0:00:22.197 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nFriday 13 March 2026  00:58:39 +0000 (0:00:00.396)       0:00:22.594 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nFriday 13 March 2026  00:58:40 +0000 (0:00:00.741)       0:00:23.335 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Include tasks from the ceph-osd role] ************************************\nFriday 13 March 2026  00:58:42 +0000 (0:00:01.773)       0:00:25.108 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m\n\nTASK [create openstack pool(s)] ************************************************\nFriday 13 March 2026  00:58:42 +0000 (0:00:00.323)       0:00:25.432 ********** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [generate keys] ***********************************************************\nFriday 13 March 2026  00:59:28 +0000 (0:00:46.397)       0:01:11.830 ********** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [get keys from monitors] **************************************************\nFriday 13 March 2026  00:59:52 +0000 (0:00:23.640)       0:01:35.470 ********** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [copy ceph key(s) if needed] **********************************************\nFriday 13 March 2026  01:00:04 +0000 (0:00:11.858)       0:01:47.329 ********** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:00:20 +0000 (0:00:16.181)       0:02:03.510 ********** \n=============================================================================== \ncreate openstack pool(s) ----------------------------------------------- 46.40s\ngenerate keys ---------------------------------------------------------- 23.64s\ncopy ceph key(s) if needed --------------------------------------------- 16.18s\nget keys from monitors ------------------------------------------------- 11.86s\nceph-facts : Get current fsid if cluster is already running ------------- 1.89s\nceph-facts : Find a running mon container ------------------------------- 1.81s\nceph-facts : Set_fact ceph_admin_command -------------------------------- 1.77s\nceph-facts : Check if the ceph mon socket is in-use --------------------- 0.86s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.81s\nceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.74s\nceph-facts : Check if podman binary is present -------------------------- 0.69s\nceph-facts : Check if the ceph conf exists ------------------------------ 0.62s\nceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.58s\nceph-facts : Check if it is atomic host --------------------------------- 0.57s\nceph-facts : Collect existed devices ------------------------------------ 0.56s\nceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.56s\nceph-facts : Read osd pool default crush rule --------------------------- 0.55s\nceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.54s\nceph-facts : Include facts.yml ------------------------------------------ 0.53s\nceph-facts : Set_fact _radosgw_address to radosgw_address --------------- 0.48s\n\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state SUCCESS\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:59:27 +0000 (0:00:00.192)       0:00:00.192 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:59:27 +0000 (0:00:00.211)       0:00:00.403 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m\n\nPLAY [Apply role horizon] ******************************************************\n\nTASK [horizon : include_tasks] *************************************************\nFriday 13 March 2026  00:59:28 +0000 (0:00:00.368)       0:00:00.772 ********** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Ensuring config directories exist] *****************************\nFriday 13 March 2026  00:59:28 +0000 (0:00:00.402)       0:00:01.174 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : Set empty custom policy] ***************************************\nFriday 13 March 2026  00:59:29 +0000 (0:00:01.132)       0:00:02.307 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.344)       0:00:02.651 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.598)       0:00:03.249 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.260)       0:00:03.509 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.094)       0:00:03.603 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:31 +0000 (0:00:00.340)       0:00:03.944 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:31 +0000 (0:00:00.261)       0:00:04.205 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:31 +0000 (0:00:00.117)       0:00:04.322 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:31 +0000 (0:00:00.242)       0:00:04.564 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:32 +0000 (0:00:00.273)       0:00:04.838 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:32 +0000 (0:00:00.224)       0:00:05.063 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:32 +0000 (0:00:00.248)       0:00:05.311 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:32 +0000 (0:00:00.257)       0:00:05.569 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:33 +0000 (0:00:00.114)       0:00:05.683 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:33 +0000 (0:00:00.232)       0:00:05.916 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:33 +0000 (0:00:00.448)       0:00:06.365 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:33 +0000 (0:00:00.122)       0:00:06.488 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:34 +0000 (0:00:00.299)       0:00:06.788 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:34 +0000 (0:00:00.305)       0:00:07.093 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:34 +0000 (0:00:00.171)       0:00:07.265 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:34 +0000 (0:00:00.262)       0:00:07.527 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:35 +0000 (0:00:00.536)       0:00:08.064 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:35 +0000 (0:00:00.175)       0:00:08.239 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:35 +0000 (0:00:00.339)       0:00:08.578 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:36 +0000 (0:00:00.306)       0:00:08.884 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:36 +0000 (0:00:00.128)       0:00:09.013 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:36 +0000 (0:00:00.458)       0:00:09.471 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:37 +0000 (0:00:00.340)       0:00:09.811 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:37 +0000 (0:00:00.114)       0:00:09.926 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nFriday 13 March 2026  00:59:37 +0000 (0:00:00.294)       0:00:10.221 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  00:59:37 +0000 (0:00:00.290)       0:00:10.511 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nFriday 13 March 2026  00:59:37 +0000 (0:00:00.122)       0:00:10.633 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over config.json files for services] *******************\nFriday 13 March 2026  00:59:38 +0000 (0:00:00.476)       0:00:11.110 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over horizon.conf] *************************************\nFriday 13 March 2026  00:59:40 +0000 (0:00:01.743)       0:00:12.854 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\nTASK [horizon : Copying over kolla-settings.py] ********************************\nFriday 13 March 2026  00:59:41 +0000 (0:00:01.638)       0:00:14.493 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over custom-settings.py] *******************************\nFriday 13 March 2026  00:59:44 +0000 (0:00:02.339)       0:00:16.832 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over existing policy file] *****************************\nFriday 13 March 2026  00:59:46 +0000 (0:00:01.960)       0:00:18.792 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over custom themes] ************************************\nFriday 13 March 2026  00:59:46 +0000 (0:00:00.291)       0:00:19.084 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nFriday 13 March 2026  00:59:46 +0000 (0:00:00.299)       0:00:19.384 ********** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over extra CA certificates] ********\nFriday 13 March 2026  00:59:47 +0000 (0:00:00.757)       0:00:20.141 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:59:49 +0000 (0:00:01.554)       0:00:21.695 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****\nFriday 13 March 2026  00:59:49 +0000 (0:00:00.622)       0:00:22.317 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : horizon | Check containers] *******************\nFriday 13 March 2026  00:59:50 +0000 (0:00:00.958)       0:00:23.275 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-check-containers : horizon | Notify handlers to restart containers] ***\nFriday 13 March 2026  00:59:52 +0000 (0:00:01.425)       0:00:24.701 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  00:59:52 +0000 (0:00:00.302)       0:00:25.004 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nFriday 13 March 2026  00:59:53 +0000 (0:00:00.865)       0:00:25.869 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nFriday 13 March 2026  00:59:53 +0000 (0:00:00.407)       0:00:26.276 ********** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Creating Horizon database] *************************************\nFriday 13 March 2026  00:59:54 +0000 (0:00:00.465)       0:00:26.741 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Creating Horizon database user and setting permissions] ********\nFriday 13 March 2026  00:59:56 +0000 (0:00:02.800)       0:00:29.542 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Running Horizon bootstrap container] ***************************\nFriday 13 March 2026  00:59:59 +0000 (0:00:02.595)       0:00:32.138 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Flush handlers] ************************************************\nFriday 13 March 2026  01:00:14 +0000 (0:00:14.887)       0:00:47.026 ********** \n\nTASK [horizon : Flush handlers] ************************************************\nFriday 13 March 2026  01:00:14 +0000 (0:00:00.077)       0:00:47.104 ********** \n\nTASK [horizon : Flush handlers] ************************************************\nFriday 13 March 2026  01:00:14 +0000 (0:00:00.244)       0:00:47.349 ********** \n\nRUNNING HANDLER [horizon : Restart horizon container] **************************\nFriday 13 March 2026  01:00:14 +0000 (0:00:00.064)       0:00:47.413 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=26  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=16  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=16  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:00:55 +0000 (0:00:41.041)       0:01:28.455 ********** \n=============================================================================== \nhorizon : Restart horizon container ------------------------------------ 41.04s\nhorizon : Running Horizon bootstrap container -------------------------- 14.89s\nhorizon : Creating Horizon database ------------------------------------- 2.80s\nhorizon : Creating Horizon database user and setting permissions -------- 2.60s\nhorizon : Copying over kolla-settings.py -------------------------------- 2.34s\nhorizon : Copying over custom-settings.py ------------------------------- 1.96s\nhorizon : Copying over config.json files for services ------------------- 1.74s\nhorizon : Copying over horizon.conf ------------------------------------- 1.64s\nservice-cert-copy : horizon | Copying over extra CA certificates -------- 1.55s\nservice-check-containers : horizon | Check containers ------------------- 1.43s\nhorizon : Ensuring config directories exist ----------------------------- 1.13s\nservice-cert-copy : horizon | Copying over backend internal TLS key ----- 0.96s\nservice-check-containers : Include tasks -------------------------------- 0.87s\nhorizon : include_tasks ------------------------------------------------- 0.76s\nservice-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.62s\nhorizon : include_tasks ------------------------------------------------- 0.60s\nhorizon : Update policy file name --------------------------------------- 0.54s\nhorizon : Update custom policy file name -------------------------------- 0.48s\nhorizon : include_tasks ------------------------------------------------- 0.47s\nhorizon : Update custom policy file name -------------------------------- 0.46s\n\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy ceph keys to the configuration repository] **************************\n\nTASK [Check if ceph keys exist] ************************************************\nFriday 13 March 2026  01:00:25 +0000 (0:00:00.141)       0:00:00.141 ********** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Fetch all ceph keys] *****************************************************\nFriday 13 March 2026  01:00:29 +0000 (0:00:04.100)       0:00:04.241 ********** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Create share directory] **************************************************\nFriday 13 March 2026  01:00:32 +0000 (0:00:03.523)       0:00:07.765 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Write ceph keys to the share directory] **********************************\nFriday 13 March 2026  01:00:33 +0000 (0:00:00.919)       0:00:08.684 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Check if target directories exist] ***************************************\nFriday 13 March 2026  01:00:46 +0000 (0:00:12.684)       0:00:21.369 ********** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m\n\nTASK [Write ceph keys to the configuration directory] **************************\nFriday 13 March 2026  01:00:49 +0000 (0:00:02.700)       0:00:24.069 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:00:54 +0000 (0:00:05.838)       0:00:29.908 ********** \n=============================================================================== \nWrite ceph keys to the share directory --------------------------------- 12.68s\nWrite ceph keys to the configuration directory -------------------------- 5.84s\nCheck if ceph keys exist ------------------------------------------------ 4.10s\nFetch all ceph keys ----------------------------------------------------- 3.52s\nCheck if target directories exist --------------------------------------- 2.70s\nCreate share directory -------------------------------------------------- 0.92s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  00:59:27 +0000 (0:00:00.214)       0:00:00.214 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  00:59:27 +0000 (0:00:00.211)       0:00:00.426 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Apply role keystone] *****************************************************\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  00:59:28 +0000 (0:00:00.360)       0:00:00.786 ********** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Ensuring config directories exist] ****************************\nFriday 13 March 2026  00:59:28 +0000 (0:00:00.431)       0:00:01.217 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Check if policies shall be overwritten] ***********************\nFriday 13 March 2026  00:59:30 +0000 (0:00:01.940)       0:00:03.158 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Set keystone policy file] *************************************\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.126)       0:00:03.284 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Check if Keystone domain-specific config is supplied] *********\nFriday 13 March 2026  00:59:30 +0000 (0:00:00.342)       0:00:03.626 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  00:59:31 +0000 (0:00:00.692)       0:00:04.319 ********** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over extra CA certificates] *******\nFriday 13 March 2026  00:59:32 +0000 (0:00:00.401)       0:00:04.721 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  00:59:35 +0000 (0:00:03.294)       0:00:08.015 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****\nFriday 13 March 2026  00:59:36 +0000 (0:00:00.619)       0:00:08.635 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over config.json files for services] ******************\nFriday 13 March 2026  00:59:36 +0000 (0:00:00.757)       0:00:09.393 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying over keystone.conf] ***********************************\nFriday 13 March 2026  00:59:40 +0000 (0:00:03.852)       0:00:13.246 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying keystone-startup script for keystone] *****************\nFriday 13 March 2026  00:59:45 +0000 (0:00:04.952)       0:00:18.198 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Create Keystone domain-specific config directory] *************\nFriday 13 March 2026  00:59:47 +0000 (0:00:01.563)       0:00:19.761 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Get file list in custom domains folder] ***********************\nFriday 13 March 2026  00:59:47 +0000 (0:00:00.539)       0:00:20.301 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying Keystone Domain specific settings] ********************\nFriday 13 March 2026  00:59:47 +0000 (0:00:00.325)       0:00:20.627 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over existing policy file] ****************************\nFriday 13 March 2026  00:59:48 +0000 (0:00:00.486)       0:00:21.113 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  00:59:49 +0000 (0:00:00.580)       0:00:21.694 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over wsgi-keystone.conf] ******************************\nFriday 13 March 2026  00:59:49 +0000 (0:00:00.301)       0:00:21.995 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\nTASK [keystone : Checking whether keystone-paste.ini file exists] **************\nFriday 13 March 2026  00:59:51 +0000 (0:00:01.671)       0:00:23.667 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Copying over keystone-paste.ini] ******************************\nFriday 13 March 2026  00:59:51 +0000 (0:00:00.872)       0:00:24.539 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Generate the required cron jobs for the node] *****************\nFriday 13 March 2026  00:59:52 +0000 (0:00:00.633)       0:00:25.172 ********** \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***\nFriday 13 March 2026  00:59:53 +0000 (0:00:01.073)       0:00:26.246 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying files for keystone-fernet] ****************************\nFriday 13 March 2026  00:59:53 +0000 (0:00:00.273)       0:00:26.520 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [keystone : Copying files for keystone-ssh] *******************************\nFriday 13 March 2026  01:00:02 +0000 (0:00:08.695)       0:00:35.216 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\nTASK [service-check-containers : keystone | Check containers] ******************\nFriday 13 March 2026  01:00:05 +0000 (0:00:02.664)       0:00:37.880 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : keystone | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:00:07 +0000 (0:00:02.422)       0:00:40.303 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:00:08 +0000 (0:00:00.360)       0:00:40.663 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  01:00:08 +0000 (0:00:00.917)       0:00:41.581 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Creating keystone database] ***********************************\nFriday 13 March 2026  01:00:09 +0000 (0:00:00.288)       0:00:41.869 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating Keystone database user and setting permissions] ******\nFriday 13 March 2026  01:00:11 +0000 (0:00:02.076)       0:00:43.946 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Checking for any running keystone_fernet containers] **********\nFriday 13 March 2026  01:00:13 +0000 (0:00:02.246)       0:00:46.192 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Group nodes where keystone_fernet is running] *****************\nFriday 13 March 2026  01:00:14 +0000 (0:00:00.861)       0:00:47.054 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***\nFriday 13 March 2026  01:00:14 +0000 (0:00:00.323)       0:00:47.378 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Running Keystone bootstrap container] *************************\nFriday 13 March 2026  01:00:15 +0000 (0:00:00.681)       0:00:48.059 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Running Keystone fernet bootstrap container] ******************\nFriday 13 March 2026  01:00:29 +0000 (0:00:13.723)       0:01:01.783 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Flush handlers] ***********************************************\nFriday 13 March 2026  01:00:38 +0000 (0:00:09.136)       0:01:10.919 ********** \n\nTASK [keystone : Flush handlers] ***********************************************\nFriday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:10.970 ********** \n\nTASK [keystone : Flush handlers] ***********************************************\nFriday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:11.020 ********** \n\nRUNNING HANDLER [keystone : Restart keystone-ssh container] ********************\nFriday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:11.071 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone-fernet container] *****************\nFriday 13 March 2026  01:00:51 +0000 (0:00:13.421)       0:01:24.492 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone container] ************************\nFriday 13 March 2026  01:01:01 +0000 (0:00:09.295)       0:01:33.788 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  01:01:07 +0000 (0:00:06.275)       0:01:40.063 ********** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Waiting for Keystone SSH port to be UP] ***********************\nFriday 13 March 2026  01:01:07 +0000 (0:00:00.505)       0:01:40.568 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Run key distribution] *****************************************\nFriday 13 March 2026  01:01:08 +0000 (0:00:00.910)       0:01:41.478 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating admin project, user, role, service, and endpoint] ****\nFriday 13 March 2026  01:01:10 +0000 (0:00:01.481)       0:01:42.960 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating/deleting services] *************\nFriday 13 March 2026  01:01:21 +0000 (0:00:11.495)       0:01:54.456 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m\n\nTASK [service-ks-register : keystone | Creating/deleting endpoints] ************\nFriday 13 March 2026  01:01:25 +0000 (0:00:04.152)       0:01:58.608 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating projects] **********************\nFriday 13 March 2026  01:01:32 +0000 (0:00:07.021)       0:02:05.629 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating users] *************************\nFriday 13 March 2026  01:01:33 +0000 (0:00:00.113)       0:02:05.742 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating roles] *************************\nFriday 13 March 2026  01:01:33 +0000 (0:00:00.111)       0:02:05.853 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Granting/revoking user roles] ***********\nFriday 13 March 2026  01:01:33 +0000 (0:00:00.115)       0:02:05.969 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating default user role] ***********************************\nFriday 13 March 2026  01:01:33 +0000 (0:00:00.302)       0:02:06.271 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nFriday 13 March 2026  01:01:36 +0000 (0:00:03.214)       0:02:09.486 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:01:37 +0000 (0:00:00.366)       0:02:09.852 ********** \n=============================================================================== \nkeystone : Running Keystone bootstrap container ------------------------ 13.72s\nkeystone : Restart keystone-ssh container ------------------------------ 13.42s\nkeystone : Creating admin project, user, role, service, and endpoint --- 11.50s\nkeystone : Restart keystone-fernet container ---------------------------- 9.30s\nkeystone : Running Keystone fernet bootstrap container ------------------ 9.14s\nkeystone : Copying files for keystone-fernet ---------------------------- 8.70s\nservice-ks-register : keystone | Creating/deleting endpoints ------------ 7.02s\nkeystone : Restart keystone container ----------------------------------- 6.28s\nkeystone : Copying over keystone.conf ----------------------------------- 4.95s\nservice-ks-register : keystone | Creating/deleting services ------------- 4.15s\nkeystone : Copying over config.json files for services ------------------ 3.85s\nservice-cert-copy : keystone | Copying over extra CA certificates ------- 3.29s\nkeystone : Creating default user role ----------------------------------- 3.21s\nkeystone : Copying files for keystone-ssh ------------------------------- 2.66s\nservice-check-containers : keystone | Check containers ------------------ 2.42s\nkeystone : Creating Keystone database user and setting permissions ------ 2.25s\nkeystone : Creating keystone database ----------------------------------- 2.08s\nkeystone : Ensuring config directories exist ---------------------------- 1.94s\nkeystone : Copying over wsgi-keystone.conf ------------------------------ 1.67s\nkeystone : Copying keystone-startup script for keystone ----------------- 1.56s\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nFriday 13 March 2026  01:00:59 +0000 (0:00:00.217)       0:00:00.217 ********** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nFriday 13 March 2026  01:00:59 +0000 (0:00:00.209)       0:00:00.427 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nFriday 13 March 2026  01:01:00 +0000 (0:00:01.090)       0:00:01.517 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nFriday 13 March 2026  01:01:01 +0000 (0:00:01.160)       0:00:02.677 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nFriday 13 March 2026  01:01:02 +0000 (0:00:00.708)       0:00:03.385 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nFriday 13 March 2026  01:01:03 +0000 (0:00:00.769)       0:00:04.155 ********** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nFriday 13 March 2026  01:01:39 +0000 (0:00:36.490)       0:00:40.645 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nFriday 13 March 2026  01:01:43 +0000 (0:00:04.228)       0:00:44.873 ********** \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nFriday 13 March 2026  01:01:44 +0000 (0:00:00.358)       0:00:45.232 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nFriday 13 March 2026  01:01:44 +0000 (0:00:00.095)       0:00:45.328 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******\nFriday 13 March 2026  01:01:44 +0000 (0:00:00.364)       0:00:45.693 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***\nFriday 13 March 2026  01:01:45 +0000 (0:00:01.127)       0:00:46.821 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******\nFriday 13 March 2026  01:01:46 +0000 (0:00:00.548)       0:00:47.369 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****\nFriday 13 March 2026  01:01:46 +0000 (0:00:00.544)       0:00:47.914 ********** \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:01:47 +0000 (0:00:01.177)       0:00:49.092 ********** \n=============================================================================== \nosism.services.cephclient : Manage cephclient service ------------------ 36.49s\nosism.services.cephclient : Copy wrapper scripts ------------------------ 4.23s\nosism.services.cephclient : Copy bash completion scripts ---------------- 1.18s\nosism.services.cephclient : Copy configuration files -------------------- 1.16s\nosism.services.cephclient : Restart cephclient service ------------------ 1.13s\nosism.services.cephclient : Create required directories ----------------- 1.09s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 0.77s\nosism.services.cephclient : Copy keyring file --------------------------- 0.71s\nosism.services.cephclient : Ensure that all containers are up ----------- 0.55s\nosism.services.cephclient : Wait for an healthy service ----------------- 0.54s\nosism.services.cephclient : Include rook task --------------------------- 0.36s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.36s\nosism.services.cephclient : Include container tasks --------------------- 0.21s\nosism.services.cephclient : Include package tasks ----------------------- 0.10s\n\n\nPLAY [Download ironic ipa images] **********************************************\n\nTASK [Ensure the destination directory exists] *********************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.115)       0:00:00.115 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent initramfs] *****************************************\nFriday 13 March 2026  01:01:42 +0000 (0:00:01.068)       0:00:01.183 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent kernel] ********************************************\nFriday 13 March 2026  01:02:12 +0000 (0:00:29.782)       0:00:30.966 ********** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (3 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (2 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:03:03 +0000 (0:00:50.676)       0:01:21.642 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:03:03 +0000 (0:00:00.281)       0:01:21.924 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m\n\nPLAY [Apply role ironic] *******************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:03:04 +0000 (0:00:01.068)       0:01:22.992 ********** \n=============================================================================== \nDownload ironic-agent kernel ------------------------------------------- 50.68s\nDownload ironic-agent initramfs ---------------------------------------- 29.78s\nGroup hosts based on enabled services ----------------------------------- 1.07s\nEnsure the destination directory exists --------------------------------- 1.07s\nGroup hosts based on Kolla action --------------------------------------- 0.28s\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Bootstraph ceph dashboard] ***********************************************\n\nTASK [Disable the ceph dashboard] **********************************************\nFriday 13 March 2026  01:01:52 +0000 (0:00:00.264)       0:00:00.264 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/ssl to false] ******************************************\nFriday 13 March 2026  01:01:53 +0000 (0:00:01.547)       0:00:01.811 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_port to 7000] ***********************************\nFriday 13 March 2026  01:01:54 +0000 (0:00:00.952)       0:00:02.764 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************\nFriday 13 March 2026  01:01:55 +0000 (0:00:00.953)       0:00:03.718 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_behaviour to error] ****************************\nFriday 13 March 2026  01:01:56 +0000 (0:00:01.025)       0:00:04.743 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_error_status_code to 404] **********************\nFriday 13 March 2026  01:01:57 +0000 (0:00:00.943)       0:00:05.687 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable the ceph dashboard] ***********************************************\nFriday 13 March 2026  01:01:58 +0000 (0:00:00.970)       0:00:06.657 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Write ceph_dashboard_password to temporary file] *************************\nFriday 13 March 2026  01:02:00 +0000 (0:00:01.196)       0:00:07.854 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Create admin user] *******************************************************\nFriday 13 March 2026  01:02:01 +0000 (0:00:01.071)       0:00:08.925 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Remove temporary file for ceph_dashboard_password] ***********************\nFriday 13 March 2026  01:02:56 +0000 (0:00:55.493)       0:01:04.418 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nFriday 13 March 2026  01:02:56 +0000 (0:00:00.168)       0:01:04.587 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nFriday 13 March 2026  01:02:58 +0000 (0:00:01.328)       0:01:05.916 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nFriday 13 March 2026  01:03:09 +0000 (0:00:11.596)       0:01:17.512 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:03:20 +0000 (0:00:11.218)       0:01:28.731 ********** \n=============================================================================== \nCreate admin user ------------------------------------------------------ 55.49s\nRestart ceph manager service ------------------------------------------- 24.14s\nDisable the ceph dashboard ---------------------------------------------- 1.55s\nEnable the ceph dashboard ----------------------------------------------- 1.20s\nWrite ceph_dashboard_password to temporary file ------------------------- 1.07s\nSet mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.03s\nSet mgr/dashboard/standby_error_status_code to 404 ---------------------- 0.97s\nSet mgr/dashboard/server_port to 7000 ----------------------------------- 0.95s\nSet mgr/dashboard/ssl to false ------------------------------------------ 0.95s\nSet mgr/dashboard/standby_behaviour to error ---------------------------- 0.94s\nRemove temporary file for ceph_dashboard_password ----------------------- 0.17s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.198)       0:00:00.198 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.220)       0:00:00.418 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m\n\nPLAY [Apply role barbican] *****************************************************\n\nTASK [barbican : include_tasks] ************************************************\nFriday 13 March 2026  01:01:42 +0000 (0:00:00.644)       0:00:01.062 ********** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : barbican | Creating/deleting services] *************\nFriday 13 March 2026  01:01:43 +0000 (0:00:00.513)       0:00:01.576 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m\n\nTASK [service-ks-register : barbican | Creating/deleting endpoints] ************\nFriday 13 March 2026  01:01:47 +0000 (0:00:04.662)       0:00:06.239 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating projects] **********************\nFriday 13 March 2026  01:01:54 +0000 (0:00:06.502)       0:00:12.741 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating users] *************************\nFriday 13 March 2026  01:01:58 +0000 (0:00:03.863)       0:00:16.605 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : barbican | Creating roles] *************************\nFriday 13 March 2026  01:02:02 +0000 (0:00:04.254)       0:00:20.859 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m\n\nTASK [service-ks-register : barbican | Granting/revoking user roles] ***********\nFriday 13 March 2026  01:02:17 +0000 (0:00:14.833)       0:00:35.692 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m\n\nTASK [barbican : Ensuring config directories exist] ****************************\nFriday 13 March 2026  01:02:21 +0000 (0:00:04.336)       0:00:40.029 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Ensuring vassals config directories exist] ********************\nFriday 13 March 2026  01:02:23 +0000 (0:00:02.269)       0:00:42.298 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m\n\nTASK [barbican : Check if policies shall be overwritten] ***********************\nFriday 13 March 2026  01:02:25 +0000 (0:00:01.257)       0:00:43.555 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Set barbican policy file] *************************************\nFriday 13 March 2026  01:02:25 +0000 (0:00:00.108)       0:00:43.664 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nFriday 13 March 2026  01:02:25 +0000 (0:00:00.396)       0:00:44.061 ********** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over extra CA certificates] *******\nFriday 13 March 2026  01:02:26 +0000 (0:00:00.460)       0:00:44.521 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:02:29 +0000 (0:00:03.130)       0:00:47.651 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****\nFriday 13 March 2026  01:02:30 +0000 (0:00:01.220)       0:00:48.872 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over config.json files for services] ******************\nFriday 13 March 2026  01:02:31 +0000 (0:00:01.339)       0:00:50.218 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over barbican-api.ini] ********************************\nFriday 13 March 2026  01:02:35 +0000 (0:00:03.630)       0:00:53.849 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [barbican : Checking whether barbican-api-paste.ini file exists] **********\nFriday 13 March 2026  01:02:38 +0000 (0:00:02.887)       0:00:56.737 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [barbican : Copying over barbican-api-paste.ini] **************************\nFriday 13 March 2026  01:02:39 +0000 (0:00:01.179)       0:00:57.916 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [barbican : Copying over barbican.conf] ***********************************\nFriday 13 March 2026  01:02:40 +0000 (0:00:00.580)       0:00:58.496 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over existing policy file] ****************************\nFriday 13 March 2026  01:02:49 +0000 (0:00:09.058)       0:01:07.555 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : barbican | Check containers] ******************\nFriday 13 March 2026  01:02:50 +0000 (0:00:01.449)       0:01:09.004 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : barbican | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:02:54 +0000 (0:00:04.118)       0:01:13.123 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:02:55 +0000 (0:00:00.857)       0:01:13.981 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nFriday 13 March 2026  01:02:56 +0000 (0:00:00.811)       0:01:14.793 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Creating barbican database] ***********************************\nFriday 13 March 2026  01:02:57 +0000 (0:00:00.995)       0:01:15.789 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Creating barbican database user and setting permissions] ******\nFriday 13 March 2026  01:02:59 +0000 (0:00:02.364)       0:01:18.154 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Running barbican bootstrap container] *************************\nFriday 13 March 2026  01:03:01 +0000 (0:00:02.044)       0:01:20.199 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Flush handlers] ***********************************************\nFriday 13 March 2026  01:03:13 +0000 (0:00:11.456)       0:01:31.655 ********** \n\nTASK [barbican : Flush handlers] ***********************************************\nFriday 13 March 2026  01:03:13 +0000 (0:00:00.144)       0:01:31.801 ********** \n\nTASK [barbican : Flush handlers] ***********************************************\nFriday 13 March 2026  01:03:13 +0000 (0:00:00.103)       0:01:31.904 ********** \n\nRUNNING HANDLER [barbican : Restart barbican-api container] ********************\nFriday 13 March 2026  01:03:13 +0000 (0:00:00.142)       0:01:32.047 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******\nFriday 13 March 2026  01:03:24 +0000 (0:00:11.036)       0:01:43.083 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-worker container] *****************\nFriday 13 March 2026  01:03:29 +0000 (0:00:04.877)       0:01:47.961 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:03:34 +0000 (0:00:05.069)       0:01:53.030 ********** \n=============================================================================== \nservice-ks-register : barbican | Creating roles ------------------------ 14.83s\nbarbican : Running barbican bootstrap container ------------------------ 11.46s\nbarbican : Restart barbican-api container ------------------------------ 11.04s\nbarbican : Copying over barbican.conf ----------------------------------- 9.06s\nservice-ks-register : barbican | Creating/deleting endpoints ------------ 6.50s\nbarbican : Restart barbican-worker container ---------------------------- 5.07s\nbarbican : Restart barbican-keystone-listener container ----------------- 4.88s\nservice-ks-register : barbican | Creating/deleting services ------------- 4.66s\nservice-ks-register : barbican | Granting/revoking user roles ----------- 4.34s\nservice-ks-register : barbican | Creating users ------------------------- 4.25s\nservice-check-containers : barbican | Check containers ------------------ 4.12s\nservice-ks-register : barbican | Creating projects ---------------------- 3.86s\nbarbican : Copying over config.json files for services ------------------ 3.63s\nservice-cert-copy : barbican | Copying over extra CA certificates ------- 3.13s\nbarbican : Copying over barbican-api.ini -------------------------------- 2.89s\nbarbican : Creating barbican database ----------------------------------- 2.36s\nbarbican : Ensuring config directories exist ---------------------------- 2.27s\nbarbican : Creating barbican database user and setting permissions ------ 2.05s\nbarbican : Copying over existing policy file ---------------------------- 1.45s\nservice-cert-copy : barbican | Copying over backend internal TLS key ---- 1.35s\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:03:11 +0000 (0:00:00.542)       0:00:00.542 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:03:11 +0000 (0:00:00.349)       0:00:00.892 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m\n\nPLAY [Apply role placement] ****************************************************\n\nTASK [placement : include_tasks] ***********************************************\nFriday 13 March 2026  01:03:11 +0000 (0:00:00.373)       0:00:01.265 ********** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : placement | Creating/deleting services] ************\nFriday 13 March 2026  01:03:12 +0000 (0:00:00.486)       0:00:01.752 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m\n\nTASK [service-ks-register : placement | Creating/deleting endpoints] ***********\nFriday 13 March 2026  01:03:15 +0000 (0:00:03.256)       0:00:05.008 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m\n\nTASK [service-ks-register : placement | Creating projects] *********************\nFriday 13 March 2026  01:03:21 +0000 (0:00:06.138)       0:00:11.147 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating users] ************************\nFriday 13 March 2026  01:03:24 +0000 (0:00:03.327)       0:00:14.475 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : placement | Creating roles] ************************\nFriday 13 March 2026  01:03:28 +0000 (0:00:03.632)       0:00:18.107 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : placement | Granting/revoking user roles] **********\nFriday 13 March 2026  01:03:31 +0000 (0:00:03.332)       0:00:21.439 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nFriday 13 March 2026  01:03:36 +0000 (0:00:04.163)       0:00:25.603 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Ensuring config directories exist] ***************************\nFriday 13 March 2026  01:03:36 +0000 (0:00:00.540)       0:00:26.143 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [placement : Check if policies shall be overwritten] **********************\nFriday 13 March 2026  01:03:37 +0000 (0:00:00.937)       0:00:27.081 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [placement : Set placement policy file] ***********************************\nFriday 13 March 2026  01:03:37 +0000 (0:00:00.131)       0:00:27.213 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nFriday 13 March 2026  01:03:38 +0000 (0:00:00.753)       0:00:27.966 ********** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over extra CA certificates] ******\nFriday 13 March 2026  01:03:39 +0000 (0:00:00.825)       0:00:28.791 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:03:40 +0000 (0:00:01.535)       0:00:30.327 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS key] ***\nFriday 13 March 2026  01:03:41 +0000 (0:00:01.007)       0:00:31.335 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over config.json files for services] *****************\nFriday 13 March 2026  01:03:43 +0000 (0:00:01.209)       0:00:32.544 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [placement : Copying over placement.conf] *********************************\nFriday 13 March 2026  01:03:44 +0000 (0:00:01.781)       0:00:34.325 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [placement : Copying over placement-api wsgi configuration] ***************\nFriday 13 March 2026  01:03:48 +0000 (0:00:03.272)       0:00:37.597 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Configure uWSGI for Placement] *******************************************\nFriday 13 March 2026  01:03:48 +0000 (0:00:00.409)       0:00:38.007 ********** \n\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-uwsgi-config : Copying over placement-api uWSGI config] **********\nFriday 13 March 2026  01:03:49 +0000 (0:00:01.130)       0:00:39.137 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [placement : Copying over migrate-db.rc.j2 configuration] *****************\nFriday 13 March 2026  01:03:51 +0000 (0:00:02.123)       0:00:41.260 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over existing policy file] ***************************\nFriday 13 March 2026  01:03:53 +0000 (0:00:01.320)       0:00:42.581 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : placement | Check containers] *****************\nFriday 13 March 2026  01:03:53 +0000 (0:00:00.580)       0:00:43.162 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m\n\nTASK [service-check-containers : placement | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:03:54 +0000 (0:00:01.192)       0:00:44.354 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:03:55 +0000 (0:00:00.460)       0:00:44.815 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Creating placement databases] ********************************\nFriday 13 March 2026  01:03:55 +0000 (0:00:00.580)       0:00:45.395 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Creating placement databases user and setting permissions] ***\nFriday 13 March 2026  01:03:58 +0000 (0:00:02.198)       0:00:47.593 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Running placement bootstrap container] ***********************\nFriday 13 March 2026  01:04:00 +0000 (0:00:02.294)       0:00:49.888 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Flush handlers] **********************************************\nFriday 13 March 2026  01:04:13 +0000 (0:00:12.858)       0:01:02.746 ********** \n\nTASK [placement : Flush handlers] **********************************************\nFriday 13 March 2026  01:04:13 +0000 (0:00:00.060)       0:01:02.807 ********** \n\nTASK [placement : Flush handlers] **********************************************\nFriday 13 March 2026  01:04:13 +0000 (0:00:00.249)       0:01:03.056 ********** \n\nRUNNING HANDLER [placement : Restart placement-api container] ******************\nFriday 13 March 2026  01:04:13 +0000 (0:00:00.071)       0:01:03.128 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:04:18 +0000 (0:00:04.385)       0:01:07.513 ********** \n=============================================================================== \nplacement : Running placement bootstrap container ---------------------- 12.86s\nservice-ks-register : placement | Creating/deleting endpoints ----------- 6.14s\nplacement : Restart placement-api container ----------------------------- 4.39s\nservice-ks-register : placement | Granting/revoking user roles ---------- 4.16s\nservice-ks-register : placement | Creating users ------------------------ 3.63s\nservice-ks-register : placement | Creating roles ------------------------ 3.33s\nservice-ks-register : placement | Creating projects --------------------- 3.33s\nplacement : Copying over placement.conf --------------------------------- 3.27s\nservice-ks-register : placement | Creating/deleting services ------------ 3.26s\nplacement : Creating placement databases user and setting permissions --- 2.29s\nplacement : Creating placement databases -------------------------------- 2.20s\nservice-uwsgi-config : Copying over placement-api uWSGI config ---------- 2.12s\nplacement : Copying over config.json files for services ----------------- 1.78s\nservice-cert-copy : placement | Copying over extra CA certificates ------ 1.54s\nplacement : Copying over migrate-db.rc.j2 configuration ----------------- 1.32s\nservice-cert-copy : placement | Copying over backend internal TLS key --- 1.21s\nservice-check-containers : placement | Check containers ----------------- 1.19s\nConfigure uWSGI for Placement ------------------------------------------- 1.13s\nservice-cert-copy : placement | Copying over backend internal TLS certificate --- 1.01s\nplacement : Ensuring config directories exist --------------------------- 0.94s\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.207)       0:00:00.207 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.237)       0:00:00.444 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m\n\nPLAY [Apply role designate] ****************************************************\n\nTASK [designate : include_tasks] ***********************************************\nFriday 13 March 2026  01:01:42 +0000 (0:00:00.349)       0:00:00.794 ********** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : designate | Creating/deleting services] ************\nFriday 13 March 2026  01:01:42 +0000 (0:00:00.652)       0:00:01.447 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m\n\nTASK [service-ks-register : designate | Creating/deleting endpoints] ***********\nFriday 13 March 2026  01:01:47 +0000 (0:00:04.703)       0:00:06.150 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m\n\nTASK [service-ks-register : designate | Creating projects] *********************\nFriday 13 March 2026  01:01:53 +0000 (0:00:06.304)       0:00:12.455 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating users] ************************\nFriday 13 March 2026  01:01:57 +0000 (0:00:03.534)       0:00:15.989 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : designate | Creating roles] ************************\nFriday 13 March 2026  01:02:01 +0000 (0:00:04.423)       0:00:20.413 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : designate | Granting/revoking user roles] **********\nFriday 13 March 2026  01:02:05 +0000 (0:00:03.373)       0:00:23.786 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m\n\nTASK [designate : Ensuring config directories exist] ***************************\nFriday 13 March 2026  01:02:09 +0000 (0:00:03.846)       0:00:27.633 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Check if policies shall be overwritten] **********************\nFriday 13 March 2026  01:02:11 +0000 (0:00:02.528)       0:00:30.161 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Set designate policy file] ***********************************\nFriday 13 March 2026  01:02:11 +0000 (0:00:00.125)       0:00:30.287 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nFriday 13 March 2026  01:02:12 +0000 (0:00:00.254)       0:00:30.542 ********** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over extra CA certificates] ******\nFriday 13 March 2026  01:02:12 +0000 (0:00:00.618)       0:00:31.160 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:02:18 +0000 (0:00:06.050)       0:00:37.211 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS key] ***\nFriday 13 March 2026  01:02:20 +0000 (0:00:01.510)       0:00:38.721 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Copying over config.json files for services] *****************\nFriday 13 March 2026  01:02:22 +0000 (0:00:02.688)       0:00:41.410 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over designate.conf] *********************************\nFriday 13 March 2026  01:02:29 +0000 (0:00:06.539)       0:00:47.949 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over pools.yaml] *************************************\nFriday 13 March 2026  01:02:47 +0000 (0:00:17.558)       0:01:05.508 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\nTASK [designate : Copying over named.conf] *************************************\nFriday 13 March 2026  01:02:53 +0000 (0:00:06.485)       0:01:11.994 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\nTASK [designate : Copying over rndc.conf] **************************************\nFriday 13 March 2026  01:02:56 +0000 (0:00:03.119)       0:01:15.113 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over rndc.key] ***************************************\nFriday 13 March 2026  01:03:00 +0000 (0:00:03.817)       0:01:18.931 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nFriday 13 March 2026  01:03:03 +0000 (0:00:02.919)       0:01:21.850 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over existing policy file] ***************************\nFriday 13 March 2026  01:03:04 +0000 (0:00:00.951)       0:01:22.802 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : designate | Check containers] *****************\nFriday 13 March 2026  01:03:05 +0000 (0:00:01.239)       0:01:24.041 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : designate | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:03:11 +0000 (0:00:06.103)       0:01:30.144 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:03:11 +0000 (0:00:00.294)       0:01:30.439 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nFriday 13 March 2026  01:03:14 +0000 (0:00:02.174)       0:01:32.613 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Creating Designate databases] ********************************\nFriday 13 March 2026  01:03:14 +0000 (0:00:00.236)       0:01:32.850 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m\n\nTASK [designate : Creating Designate databases user and setting permissions] ***\nFriday 13 March 2026  01:03:16 +0000 (0:00:01.896)       0:01:34.747 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m\n\nTASK [designate : Running Designate bootstrap container] ***********************\nFriday 13 March 2026  01:03:18 +0000 (0:00:02.129)       0:01:36.877 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [designate : Flush handlers] **********************************************\nFriday 13 March 2026  01:03:32 +0000 (0:00:14.152)       0:01:51.029 ********** \n\nTASK [designate : Flush handlers] **********************************************\nFriday 13 March 2026  01:03:32 +0000 (0:00:00.050)       0:01:51.080 ********** \n\nTASK [designate : Flush handlers] **********************************************\nFriday 13 March 2026  01:03:32 +0000 (0:00:00.093)       0:01:51.174 ********** \n\nRUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********\nFriday 13 March 2026  01:03:32 +0000 (0:00:00.108)       0:01:51.283 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-api container] ******************\nFriday 13 March 2026  01:03:41 +0000 (0:00:08.494)       0:01:59.777 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-central container] **************\nFriday 13 March 2026  01:03:52 +0000 (0:00:11.136)       0:02:10.913 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-producer container] *************\nFriday 13 March 2026  01:03:57 +0000 (0:00:05.265)       0:02:16.179 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-mdns container] *****************\nFriday 13 March 2026  01:04:02 +0000 (0:00:04.621)       0:02:20.800 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-worker container] ***************\nFriday 13 March 2026  01:04:10 +0000 (0:00:08.110)       0:02:28.910 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [designate : Non-destructive DNS pools update] ****************************\nFriday 13 March 2026  01:04:15 +0000 (0:00:05.175)       0:02:34.086 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=25  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:04:22 +0000 (0:00:06.833)       0:02:40.920 ********** \n=============================================================================== \ndesignate : Copying over designate.conf -------------------------------- 17.56s\ndesignate : Running Designate bootstrap container ---------------------- 14.15s\ndesignate : Restart designate-api container ---------------------------- 11.14s\ndesignate : Restart designate-backend-bind9 container ------------------- 8.50s\ndesignate : Restart designate-mdns container ---------------------------- 8.11s\ndesignate : Non-destructive DNS pools update ---------------------------- 6.83s\ndesignate : Copying over config.json files for services ----------------- 6.54s\ndesignate : Copying over pools.yaml ------------------------------------- 6.49s\nservice-ks-register : designate | Creating/deleting endpoints ----------- 6.30s\nservice-check-containers : designate | Check containers ----------------- 6.10s\nservice-cert-copy : designate | Copying over extra CA certificates ------ 6.05s\ndesignate : Restart designate-central container ------------------------- 5.27s\ndesignate : Restart designate-worker container -------------------------- 5.18s\nservice-ks-register : designate | Creating/deleting services ------------ 4.70s\ndesignate : Restart designate-producer container ------------------------ 4.62s\nservice-ks-register : designate | Creating users ------------------------ 4.42s\nservice-ks-register : designate | Granting/revoking user roles ---------- 3.85s\ndesignate : Copying over rndc.conf -------------------------------------- 3.82s\nservice-ks-register : designate | Creating projects --------------------- 3.53s\nservice-ks-register : designate | Creating roles ------------------------ 3.37s\n\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:04:22 +0000 (0:00:00.128)       0:00:00.128 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:04:22 +0000 (0:00:00.223)       0:00:00.351 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Wait for the Keystone service] *******************************************\n\nTASK [Waiting for Keystone public port to be UP] *******************************\nFriday 13 March 2026  01:04:22 +0000 (0:00:00.614)       0:00:00.966 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:04:23 +0000 (0:00:00.665)       0:00:01.631 ********** \n=============================================================================== \nWaiting for Keystone public port to be UP ------------------------------- 0.67s\nGroup hosts based on enabled services ----------------------------------- 0.62s\nGroup hosts based on Kolla action --------------------------------------- 0.22s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:04:27 +0000 (0:00:00.247)       0:00:00.247 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:04:28 +0000 (0:00:00.728)       0:00:00.976 ********** \n\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-rgw : include_tasks] ************************************************\nFriday 13 March 2026  01:04:29 +0000 (0:00:00.871)       0:00:01.848 ********** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating/deleting services] *************\nFriday 13 March 2026  01:04:31 +0000 (0:00:02.109)       0:00:03.957 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating/deleting endpoints] ************\nFriday 13 March 2026  01:04:34 +0000 (0:00:03.136)       0:00:07.098 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating projects] **********************\nFriday 13 March 2026  01:04:40 +0000 (0:00:06.127)       0:00:13.225 ********** \n\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating users] *************************\nFriday 13 March 2026  01:04:43 +0000 (0:00:02.786)       0:00:16.012 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating roles] *************************\nFriday 13 March 2026  01:04:46 +0000 (0:00:03.058)       0:00:19.070 ********** \n\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Granting/revoking user roles] ***********\nFriday 13 March 2026  01:04:52 +0000 (0:00:05.394)       0:00:24.465 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:04:55 +0000 (0:00:03.769)       0:00:28.234 ********** \n=============================================================================== \nservice-ks-register : ceph-rgw | Creating/deleting endpoints ------------ 6.13s\nservice-ks-register : ceph-rgw | Creating roles ------------------------- 5.39s\nservice-ks-register : ceph-rgw | Granting/revoking user roles ----------- 3.77s\nservice-ks-register : ceph-rgw | Creating/deleting services ------------- 3.14s\nservice-ks-register : ceph-rgw | Creating users ------------------------- 3.06s\nservice-ks-register : ceph-rgw | Creating projects ---------------------- 2.79s\nceph-rgw : include_tasks ------------------------------------------------ 2.11s\nGroup hosts based on enabled services ----------------------------------- 0.87s\nGroup hosts based on Kolla action --------------------------------------- 0.73s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:03:40 +0000 (0:00:00.484)       0:00:00.484 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:03:41 +0000 (0:00:00.359)       0:00:00.843 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m\n\nPLAY [Apply role magnum] *******************************************************\n\nTASK [magnum : include_tasks] **************************************************\nFriday 13 March 2026  01:03:41 +0000 (0:00:00.723)       0:00:01.567 ********** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : magnum | Creating/deleting services] ***************\nFriday 13 March 2026  01:03:42 +0000 (0:00:01.004)       0:00:02.572 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m\n\nTASK [service-ks-register : magnum | Creating/deleting endpoints] **************\nFriday 13 March 2026  01:03:46 +0000 (0:00:03.435)       0:00:06.007 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating projects] ************************\nFriday 13 March 2026  01:03:53 +0000 (0:00:06.894)       0:00:12.905 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating users] ***************************\nFriday 13 March 2026  01:03:56 +0000 (0:00:03.285)       0:00:16.190 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : magnum | Creating roles] ***************************\nFriday 13 March 2026  01:04:00 +0000 (0:00:03.789)       0:00:19.980 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : magnum | Granting/revoking user roles] *************\nFriday 13 March 2026  01:04:03 +0000 (0:00:03.361)       0:00:23.342 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m\n\nTASK [magnum : Creating Magnum trustee domain] *********************************\nFriday 13 March 2026  01:04:07 +0000 (0:00:03.680)       0:00:27.023 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user] ***********************************\nFriday 13 March 2026  01:04:10 +0000 (0:00:03.025)       0:00:30.048 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user role] ******************************\nFriday 13 March 2026  01:04:13 +0000 (0:00:03.497)       0:00:33.545 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Ensuring config directories exist] ******************************\nFriday 13 March 2026  01:04:16 +0000 (0:00:03.183)       0:00:36.729 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Check if policies shall be overwritten] *************************\nFriday 13 March 2026  01:04:18 +0000 (0:00:01.208)       0:00:37.937 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Set magnum policy file] *****************************************\nFriday 13 March 2026  01:04:18 +0000 (0:00:00.141)       0:00:38.079 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check if kubeconfig file is supplied] ***************************\nFriday 13 March 2026  01:04:18 +0000 (0:00:00.400)       0:00:38.479 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [magnum : Copying over kubeconfig file] ***********************************\nFriday 13 March 2026  01:04:19 +0000 (0:00:00.803)       0:00:39.283 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Set magnum kubeconfig file's path] ******************************\nFriday 13 March 2026  01:04:21 +0000 (0:00:02.201)       0:00:41.485 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nFriday 13 March 2026  01:04:21 +0000 (0:00:00.271)       0:00:41.756 ********** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over extra CA certificates] *********\nFriday 13 March 2026  01:04:22 +0000 (0:00:00.624)       0:00:42.381 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:04:24 +0000 (0:00:02.065)       0:00:44.446 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******\nFriday 13 March 2026  01:04:25 +0000 (0:00:00.760)       0:00:45.207 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Copying over config.json files for services] ********************\nFriday 13 March 2026  01:04:26 +0000 (0:00:01.282)       0:00:46.490 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over magnum.conf] ***************************************\nFriday 13 March 2026  01:04:28 +0000 (0:00:01.972)       0:00:48.462 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over existing policy file] ******************************\nFriday 13 March 2026  01:04:34 +0000 (0:00:05.452)       0:00:53.915 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : magnum | Check containers] ********************\nFriday 13 March 2026  01:04:35 +0000 (0:00:01.101)       0:00:55.016 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : magnum | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:04:39 +0000 (0:00:03.831)       0:00:58.848 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:04:39 +0000 (0:00:00.667)       0:00:59.515 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nFriday 13 March 2026  01:04:41 +0000 (0:00:01.815)       0:01:01.330 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Creating Magnum database] ***************************************\nFriday 13 March 2026  01:04:42 +0000 (0:00:00.815)       0:01:02.146 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database user and setting permissions] **********\nFriday 13 March 2026  01:04:44 +0000 (0:00:02.072)       0:01:04.218 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Running Magnum bootstrap container] *****************************\nFriday 13 March 2026  01:04:46 +0000 (0:00:01.997)       0:01:06.216 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Flush handlers] *************************************************\nFriday 13 March 2026  01:05:01 +0000 (0:00:15.374)       0:01:21.590 ********** \n\nTASK [magnum : Flush handlers] *************************************************\nFriday 13 March 2026  01:05:01 +0000 (0:00:00.061)       0:01:21.652 ********** \n\nTASK [magnum : Flush handlers] *************************************************\nFriday 13 March 2026  01:05:01 +0000 (0:00:00.064)       0:01:21.716 ********** \n\nRUNNING HANDLER [magnum : Restart magnum-api container] ************************\nFriday 13 March 2026  01:05:01 +0000 (0:00:00.065)       0:01:21.782 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [magnum : Restart magnum-conductor container] ******************\nFriday 13 March 2026  01:05:14 +0000 (0:00:12.828)       0:01:34.610 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:05:28 +0000 (0:00:13.867)       0:01:48.477 ********** \n=============================================================================== \nmagnum : Running Magnum bootstrap container ---------------------------- 15.38s\nmagnum : Restart magnum-conductor container ---------------------------- 13.87s\nmagnum : Restart magnum-api container ---------------------------------- 12.83s\nservice-ks-register : magnum | Creating/deleting endpoints -------------- 6.90s\nmagnum : Copying over magnum.conf --------------------------------------- 5.45s\nservice-check-containers : magnum | Check containers -------------------- 3.83s\nservice-ks-register : magnum | Creating users --------------------------- 3.79s\nservice-ks-register : magnum | Granting/revoking user roles ------------- 3.68s\nmagnum : Creating Magnum trustee user ----------------------------------- 3.50s\nservice-ks-register : magnum | Creating/deleting services --------------- 3.44s\nservice-ks-register : magnum | Creating roles --------------------------- 3.36s\nservice-ks-register : magnum | Creating projects ------------------------ 3.29s\nmagnum : Creating Magnum trustee user role ------------------------------ 3.18s\nmagnum : Creating Magnum trustee domain --------------------------------- 3.03s\nmagnum : Copying over kubeconfig file ----------------------------------- 2.20s\nmagnum : Creating Magnum database --------------------------------------- 2.07s\nservice-cert-copy : magnum | Copying over extra CA certificates --------- 2.07s\nmagnum : Creating Magnum database user and setting permissions ---------- 2.00s\nmagnum : Copying over config.json files for services -------------------- 1.97s\nservice-check-containers : Include tasks -------------------------------- 1.81s\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:01:41 +0000 (0:00:00.301)       0:00:00.301 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:01:42 +0000 (0:00:00.890)       0:00:01.191 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m\n\nPLAY [Apply role neutron] ******************************************************\n\nTASK [neutron : include_tasks] *************************************************\nFriday 13 March 2026  01:01:43 +0000 (0:00:00.909)       0:00:02.101 ********** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [neutron : Get container facts] *******************************************\nFriday 13 March 2026  01:01:44 +0000 (0:00:00.902)       0:00:03.004 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Get container volume facts] ************************************\nFriday 13 March 2026  01:01:45 +0000 (0:00:01.461)       0:00:04.465 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check for ML2/OVN presence] ************************************\nFriday 13 March 2026  01:01:46 +0000 (0:00:01.188)       0:00:05.654 ********** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [neutron : Check for ML2/OVS presence] ************************************\nFriday 13 March 2026  01:01:47 +0000 (0:00:00.583)       0:00:06.237 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-ks-register : neutron | Creating/deleting services] **************\nFriday 13 March 2026  01:01:47 +0000 (0:00:00.513)       0:00:06.751 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m\n\nTASK [service-ks-register : neutron | Creating/deleting endpoints] *************\nFriday 13 March 2026  01:01:51 +0000 (0:00:03.728)       0:00:10.479 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating projects] ***********************\nFriday 13 March 2026  01:01:58 +0000 (0:00:06.687)       0:00:17.167 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating users] **************************\nFriday 13 March 2026  01:02:02 +0000 (0:00:03.776)       0:00:20.943 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : neutron | Creating roles] **************************\nFriday 13 March 2026  01:02:06 +0000 (0:00:03.998)       0:00:24.942 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : neutron | Granting/revoking user roles] ************\nFriday 13 March 2026  01:02:09 +0000 (0:00:03.541)       0:00:28.484 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nFriday 13 March 2026  01:02:16 +0000 (0:00:06.810)       0:00:35.294 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Load and persist kernel modules] *****************************************\nFriday 13 March 2026  01:02:17 +0000 (0:00:00.725)       0:00:36.020 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check IPv6 support] ********************************************\nFriday 13 March 2026  01:02:19 +0000 (0:00:01.988)       0:00:38.008 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nFriday 13 March 2026  01:02:20 +0000 (0:00:01.079)       0:00:39.087 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Ensuring config directories exist] *****************************\nFriday 13 March 2026  01:02:22 +0000 (0:00:02.605)       0:00:41.693 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Check if extra ml2 plugins exists] *****************************\nFriday 13 March 2026  01:02:25 +0000 (0:00:03.107)       0:00:44.801 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m\n\u001b[1;35ma directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nFriday 13 March 2026  01:02:26 +0000 (0:00:00.880)       0:00:45.682 ********** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over extra CA certificates] ********\nFriday 13 March 2026  01:02:27 +0000 (0:00:01.100)       0:00:46.783 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:02:30 +0000 (0:00:03.082)       0:00:49.866 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****\nFriday 13 March 2026  01:02:33 +0000 (0:00:02.769)       0:00:52.636 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Creating TLS backend PEM File] *********************************\nFriday 13 March 2026  01:02:37 +0000 (0:00:03.463)       0:00:56.099 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  01:02:39 +0000 (0:00:02.201)       0:00:58.301 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Set neutron policy file] ***************************************\nFriday 13 March 2026  01:02:39 +0000 (0:00:00.091)       0:00:58.393 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over existing policy file] *****************************\nFriday 13 March 2026  01:02:40 +0000 (0:00:00.749)       0:00:59.143 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over config.json files for services] *******************\nFriday 13 March 2026  01:02:42 +0000 (0:00:02.707)       0:01:01.851 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron.conf] *************************************\nFriday 13 March 2026  01:02:46 +0000 (0:00:03.832)       0:01:05.684 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron_vpnaas.conf] ******************************\nFriday 13 March 2026  01:02:53 +0000 (0:00:06.656)       0:01:12.340 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ssh key] ******************************************\nFriday 13 March 2026  01:02:56 +0000 (0:00:02.999)       0:01:15.340 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [neutron : Copying over ml2_conf.ini] *************************************\nFriday 13 March 2026  01:02:59 +0000 (0:00:02.948)       0:01:18.288 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [neutron : Copying over openvswitch_agent.ini] ****************************\nFriday 13 March 2026  01:03:02 +0000 (0:00:03.509)       0:01:21.798 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over sriov_agent.ini] **********************************\nFriday 13 March 2026  01:03:05 +0000 (0:00:02.736)       0:01:24.535 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over mlnx_agent.ini] ***********************************\nFriday 13 March 2026  01:03:08 +0000 (0:00:02.410)       0:01:26.945 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [neutron : Copying over eswitchd.conf] ************************************\nFriday 13 March 2026  01:03:09 +0000 (0:00:01.812)       0:01:28.758 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over dhcp_agent.ini] ***********************************\nFriday 13 March 2026  01:03:11 +0000 (0:00:01.903)       0:01:30.662 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over dnsmasq.conf] *************************************\nFriday 13 March 2026  01:03:14 +0000 (0:00:02.371)       0:01:33.033 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over l3_agent.ini] *************************************\nFriday 13 March 2026  01:03:16 +0000 (0:00:02.222)       0:01:35.256 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over fwaas_driver.ini] *********************************\nFriday 13 March 2026  01:03:17 +0000 (0:00:01.614)       0:01:36.870 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metadata_agent.ini] *******************************\nFriday 13 March 2026  01:03:19 +0000 (0:00:01.901)       0:01:38.772 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************\nFriday 13 March 2026  01:03:21 +0000 (0:00:01.686)       0:01:40.459 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over metering_agent.ini] *******************************\nFriday 13 March 2026  01:03:24 +0000 (0:00:02.934)       0:01:43.393 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ironic_neutron_agent.ini] *************************\nFriday 13 March 2026  01:03:26 +0000 (0:00:02.290)       0:01:45.684 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over bgp_dragent.ini] **********************************\nFriday 13 March 2026  01:03:28 +0000 (0:00:01.541)       0:01:47.226 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ovn_agent.ini] ************************************\nFriday 13 March 2026  01:03:30 +0000 (0:00:02.018)       0:01:49.244 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over nsx.ini] ******************************************\nFriday 13 March 2026  01:03:32 +0000 (0:00:02.047)       0:01:51.292 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copy neutron-l3-agent-wrapper script] **************************\nFriday 13 March 2026  01:03:34 +0000 (0:00:02.261)       0:01:53.554 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over extra ml2 plugins] ********************************\nFriday 13 March 2026  01:03:37 +0000 (0:00:02.601)       0:01:56.155 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************\nFriday 13 March 2026  01:03:39 +0000 (0:00:02.359)       0:01:58.514 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron_taas.conf] ********************************\nFriday 13 March 2026  01:03:41 +0000 (0:00:02.214)       0:02:00.729 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [service-check-containers : neutron | Check containers] *******************\nFriday 13 March 2026  01:03:44 +0000 (0:00:03.103)       0:02:03.832 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : neutron | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:03:47 +0000 (0:00:02.887)       0:02:06.720 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:03:48 +0000 (0:00:00.673)       0:02:07.394 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nFriday 13 March 2026  01:03:51 +0000 (0:00:02.683)       0:02:10.078 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating Neutron database] *************************************\nFriday 13 March 2026  01:03:51 +0000 (0:00:00.503)       0:02:10.581 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Creating Neutron database user and setting permissions] ********\nFriday 13 March 2026  01:03:53 +0000 (0:00:02.014)       0:02:12.595 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Running Neutron bootstrap container] ***************************\nFriday 13 March 2026  01:03:56 +0000 (0:00:02.368)       0:02:14.964 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:37.479)       0:02:52.443 ********** \n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.051)       0:02:52.495 ********** \n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.161)       0:02:52.657 ********** \n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.048)       0:02:52.705 ********** \n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.051)       0:02:52.757 ********** \n\nTASK [neutron : Flush Handlers] ************************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.047)       0:02:52.804 ********** \n\nRUNNING HANDLER [neutron : Restart neutron-server container] *******************\nFriday 13 March 2026  01:04:33 +0000 (0:00:00.048)       0:02:52.853 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******\nFriday 13 March 2026  01:05:00 +0000 (0:00:26.639)       0:03:19.493 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:05:53 +0000 (0:00:52.467)       0:04:11.961 ********** \n=============================================================================== \nneutron : Restart neutron-ovn-metadata-agent container ----------------- 52.47s\nneutron : Running Neutron bootstrap container -------------------------- 37.48s\nneutron : Restart neutron-server container ----------------------------- 26.64s\nservice-ks-register : neutron | Granting/revoking user roles ------------ 6.81s\nservice-ks-register : neutron | Creating/deleting endpoints ------------- 6.69s\nneutron : Copying over neutron.conf ------------------------------------- 6.66s\nservice-ks-register : neutron | Creating users -------------------------- 4.00s\nneutron : Copying over config.json files for services ------------------- 3.83s\nservice-ks-register : neutron | Creating projects ----------------------- 3.78s\nservice-ks-register : neutron | Creating/deleting services -------------- 3.73s\nservice-ks-register : neutron | Creating roles -------------------------- 3.54s\nneutron : Copying over ml2_conf.ini ------------------------------------- 3.51s\nservice-cert-copy : neutron | Copying over backend internal TLS key ----- 3.46s\nneutron : Ensuring config directories exist ----------------------------- 3.11s\nneutron : Copying over neutron_taas.conf -------------------------------- 3.10s\nservice-cert-copy : neutron | Copying over extra CA certificates -------- 3.08s\nneutron : Copying over neutron_vpnaas.conf ------------------------------ 3.00s\nneutron : Copying over ssh key ------------------------------------------ 2.95s\nneutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 2.93s\nservice-check-containers : neutron | Check containers ------------------- 2.89s\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:04:27 +0000 (0:00:00.366)       0:00:00.367 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:04:27 +0000 (0:00:00.736)       0:00:01.103 ********** \n\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m\n\nPLAY [Apply role prometheus] ***************************************************\n\nTASK [prometheus : include_tasks] **********************************************\nFriday 13 March 2026  01:04:28 +0000 (0:00:00.612)       0:00:01.715 ********** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [prometheus : Ensuring config directories exist] **************************\nFriday 13 March 2026  01:04:29 +0000 (0:00:01.540)       0:00:03.256 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : include_tasks] **********************************************\nFriday 13 March 2026  01:04:33 +0000 (0:00:03.150)       0:00:06.406 ********** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****\nFriday 13 March 2026  01:04:34 +0000 (0:00:01.580)       0:00:07.987 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:04:41 +0000 (0:00:07.320)       0:00:15.308 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***\nFriday 13 March 2026  01:04:44 +0000 (0:00:02.668)       0:00:17.976 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over config.json files] *****************************\nFriday 13 March 2026  01:04:47 +0000 (0:00:02.651)       0:00:20.628 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Find custom prometheus alert rules files] *******************\nFriday 13 March 2026  01:04:53 +0000 (0:00:06.094)       0:00:26.723 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom prometheus alert rules files] ***********\nFriday 13 March 2026  01:04:54 +0000 (0:00:01.086)       0:00:27.809 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Find prometheus common config overrides] ********************\nFriday 13 March 2026  01:04:55 +0000 (0:00:00.604)       0:00:28.414 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Find prometheus host config overrides] **********************\nFriday 13 March 2026  01:04:55 +0000 (0:00:00.729)       0:00:29.143 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over prometheus config file] ************************\nFriday 13 March 2026  01:04:57 +0000 (0:00:01.720)       0:00:30.864 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus web config file] ********************\nFriday 13 March 2026  01:05:14 +0000 (0:00:17.252)       0:00:48.117 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus alertmanager config file] ***********\nFriday 13 March 2026  01:05:18 +0000 (0:00:04.018)       0:00:52.135 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m\n\nTASK [prometheus : Find custom Alertmanager alert notification templates] ******\nFriday 13 March 2026  01:05:20 +0000 (0:00:01.487)       0:00:53.623 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom Alertmanager alert notification templates] ***\nFriday 13 March 2026  01:05:20 +0000 (0:00:00.664)       0:00:54.287 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************\nFriday 13 March 2026  01:05:21 +0000 (0:00:00.578)       0:00:54.866 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [prometheus : Copying cloud config file for openstack exporter] ***********\nFriday 13 March 2026  01:05:23 +0000 (0:00:01.829)       0:00:56.696 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying config file for blackbox exporter] ******************\nFriday 13 March 2026  01:05:24 +0000 (0:00:01.320)       0:00:58.016 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m\n\nTASK [prometheus : Find extra prometheus server config files] ******************\nFriday 13 March 2026  01:05:26 +0000 (0:00:01.627)       0:00:59.644 ********** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m\n\u001b[1;35mnot a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Create subdirectories for extra config files] ***************\nFriday 13 March 2026  01:05:27 +0000 (0:00:01.084)       0:01:00.729 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Template extra prometheus server config files] **************\nFriday 13 March 2026  01:05:28 +0000 (0:00:00.822)       0:01:01.551 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-check-containers : prometheus | Check containers] ****************\nFriday 13 March 2026  01:05:28 +0000 (0:00:00.595)       0:01:02.146 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-check-containers : prometheus | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:05:33 +0000 (0:00:04.718)       0:01:06.865 ********** \n\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:05:34 +0000 (0:00:00.725)       0:01:07.591 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Creating prometheus database user and setting permissions] ***\nFriday 13 March 2026  01:05:35 +0000 (0:00:01.748)       0:01:09.339 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:01.033)       0:01:10.372 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.051)       0:01:10.424 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.047)       0:01:10.472 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.048)       0:01:10.521 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.054)       0:01:10.575 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.049)       0:01:10.624 ********** \n\nTASK [prometheus : Flush handlers] *********************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.173)       0:01:10.798 ********** \n\nRUNNING HANDLER [prometheus : Restart prometheus-server container] *************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.064)       0:01:10.863 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******\nFriday 13 March 2026  01:05:57 +0000 (0:00:20.102)       0:01:30.966 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****\nFriday 13 March 2026  01:06:11 +0000 (0:00:13.596)       0:01:44.562 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***\nFriday 13 March 2026  01:06:17 +0000 (0:00:05.864)       0:01:50.426 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********\nFriday 13 March 2026  01:06:23 +0000 (0:00:06.180)       0:01:56.606 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******\nFriday 13 March 2026  01:06:36 +0000 (0:00:13.424)       0:02:10.030 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***\nFriday 13 March 2026  01:06:43 +0000 (0:00:06.849)       0:02:16.880 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***\nFriday 13 March 2026  01:06:54 +0000 (0:00:10.572)       0:02:27.453 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***\nFriday 13 March 2026  01:06:59 +0000 (0:00:05.117)       0:02:32.570 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=10  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:07:09 +0000 (0:00:09.842)       0:02:42.413 ********** \n=============================================================================== \nprometheus : Restart prometheus-server container ----------------------- 20.10s\nprometheus : Copying over prometheus config file ----------------------- 17.25s\nprometheus : Restart prometheus-node-exporter container ---------------- 13.60s\nprometheus : Restart prometheus-cadvisor container --------------------- 13.42s\nprometheus : Restart prometheus-elasticsearch-exporter container ------- 10.57s\nprometheus : Restart prometheus-libvirt-exporter container -------------- 9.84s\nservice-cert-copy : prometheus | Copying over extra CA certificates ----- 7.32s\nprometheus : Restart prometheus-alertmanager container ------------------ 6.85s\nprometheus : Restart prometheus-memcached-exporter container ------------ 6.18s\nprometheus : Copying over config.json files ----------------------------- 6.09s\nprometheus : Restart prometheus-mysqld-exporter container --------------- 5.86s\nprometheus : Restart prometheus-blackbox-exporter container ------------- 5.12s\nservice-check-containers : prometheus | Check containers ---------------- 4.72s\nprometheus : Copying over prometheus web config file -------------------- 4.02s\nprometheus : Ensuring config directories exist -------------------------- 3.15s\nservice-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.67s\nservice-cert-copy : prometheus | Copying over backend internal TLS key --- 2.65s\nprometheus : Copying over my.cnf for mysqld_exporter -------------------- 1.83s\nservice-check-containers : Include tasks -------------------------------- 1.75s\nprometheus : Find prometheus host config overrides ---------------------- 1.72s\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:05:00 +0000 (0:00:00.237)       0:00:00.237 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:05:00 +0000 (0:00:00.317)       0:00:00.555 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m\n\nPLAY [Apply role glance] *******************************************************\n\nTASK [glance : include_tasks] **************************************************\nFriday 13 March 2026  01:05:01 +0000 (0:00:00.430)       0:00:00.985 ********** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : glance | Creating/deleting services] ***************\nFriday 13 March 2026  01:05:01 +0000 (0:00:00.846)       0:00:01.835 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m\n\nTASK [service-ks-register : glance | Creating/deleting endpoints] **************\nFriday 13 March 2026  01:05:05 +0000 (0:00:03.362)       0:00:05.197 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m\n\nTASK [service-ks-register : glance | Creating projects] ************************\nFriday 13 March 2026  01:05:11 +0000 (0:00:06.279)       0:00:11.477 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating users] ***************************\nFriday 13 March 2026  01:05:15 +0000 (0:00:03.457)       0:00:14.935 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : glance | Creating roles] ***************************\nFriday 13 March 2026  01:05:18 +0000 (0:00:03.576)       0:00:18.511 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : glance | Granting/revoking user roles] *************\nFriday 13 March 2026  01:05:21 +0000 (0:00:03.140)       0:00:21.652 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m\n\nTASK [glance : Ensuring config directories exist] ******************************\nFriday 13 March 2026  01:05:25 +0000 (0:00:03.622)       0:00:25.274 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nFriday 13 March 2026  01:05:28 +0000 (0:00:03.017)       0:00:28.292 ********** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [glance : Ensuring glance service ceph config subdir exists] **************\nFriday 13 March 2026  01:05:29 +0000 (0:00:00.600)       0:00:28.892 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copy over multiple ceph configs for Glance] *********************\nFriday 13 March 2026  01:05:33 +0000 (0:00:04.741)       0:00:33.634 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\nTASK [glance : Copy over ceph Glance keyrings] *********************************\nFriday 13 March 2026  01:05:35 +0000 (0:00:01.639)       0:00:35.273 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m\n\nTASK [glance : Ensuring config directory has correct owner and permission] *****\nFriday 13 March 2026  01:05:36 +0000 (0:00:01.136)       0:00:36.409 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check if policies shall be overwritten] *************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.704)       0:00:37.114 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Set glance policy file] *****************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.119)       0:00:37.234 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nFriday 13 March 2026  01:05:37 +0000 (0:00:00.260)       0:00:37.494 ********** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over extra CA certificates] *********\nFriday 13 March 2026  01:05:38 +0000 (0:00:00.460)       0:00:37.955 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:05:41 +0000 (0:00:03.491)       0:00:41.447 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS key] ******\nFriday 13 March 2026  01:05:44 +0000 (0:00:02.705)       0:00:44.152 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating TLS backend PEM File] **********************************\nFriday 13 March 2026  01:05:47 +0000 (0:00:03.115)       0:00:47.268 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over config.json files for services] ********************\nFriday 13 March 2026  01:05:50 +0000 (0:00:03.419)       0:00:50.688 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : Copying over glance-api.conf] ***********************************\nFriday 13 March 2026  01:05:55 +0000 (0:00:04.447)       0:00:55.136 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-cache.conf for glance_api] ******************\nFriday 13 March 2026  01:06:03 +0000 (0:00:08.460)       0:01:03.596 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-image-import.conf] **************************\nFriday 13 March 2026  01:06:06 +0000 (0:00:03.144)       0:01:06.740 ********** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over property-protections-rules.conf] *******************\nFriday 13 March 2026  01:06:09 +0000 (0:00:02.564)       0:01:09.305 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over existing policy file] ******************************\nFriday 13 March 2026  01:06:13 +0000 (0:00:03.904)       0:01:13.210 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-haproxy-tls.cfg] ****************************\nFriday 13 March 2026  01:06:13 +0000 (0:00:00.286)       0:01:13.496 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostnqn' file for glance_api] ***********************\nFriday 13 March 2026  01:06:17 +0000 (0:00:03.545)       0:01:17.042 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-check-containers : glance | Check containers] ********************\nFriday 13 March 2026  01:06:22 +0000 (0:00:05.006)       0:01:22.049 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-check-containers : glance | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:06:28 +0000 (0:00:06.367)       0:01:28.416 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:06:28 +0000 (0:00:00.343)       0:01:28.759 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nFriday 13 March 2026  01:06:32 +0000 (0:00:03.436)       0:01:32.195 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating Glance database] ***************************************\nFriday 13 March 2026  01:06:32 +0000 (0:00:00.368)       0:01:32.564 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating Glance database user and setting permissions] **********\nFriday 13 March 2026  01:06:34 +0000 (0:00:01.697)       0:01:34.262 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Enable log_bin_trust_function_creators function] ****************\nFriday 13 March 2026  01:06:36 +0000 (0:00:01.830)       0:01:36.093 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Running Glance bootstrap container] *****************************\nFriday 13 March 2026  01:06:38 +0000 (0:00:02.016)       0:01:38.109 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Disable log_bin_trust_function_creators function] ***************\nFriday 13 March 2026  01:07:04 +0000 (0:00:25.870)       0:02:03.980 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:06 +0000 (0:00:01.956)       0:02:05.936 ********** \n\nTASK [glance : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:06 +0000 (0:00:00.059)       0:02:05.995 ********** \n\nTASK [glance : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:06 +0000 (0:00:00.065)       0:02:06.061 ********** \n\nRUNNING HANDLER [glance : Restart glance-api container] ************************\nFriday 13 March 2026  01:07:06 +0000 (0:00:00.079)       0:02:06.140 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=28  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:07:31 +0000 (0:00:24.834)       0:02:30.975 ********** \n=============================================================================== \nglance : Running Glance bootstrap container ---------------------------- 25.87s\nglance : Restart glance-api container ---------------------------------- 24.83s\nglance : Copying over glance-api.conf ----------------------------------- 8.46s\nservice-check-containers : glance | Check containers -------------------- 6.37s\nservice-ks-register : glance | Creating/deleting endpoints -------------- 6.28s\nglance : Generating 'hostnqn' file for glance_api ----------------------- 5.01s\nglance : Ensuring glance service ceph config subdir exists -------------- 4.74s\nglance : Copying over config.json files for services -------------------- 4.45s\nglance : Copying over property-protections-rules.conf ------------------- 3.90s\nservice-ks-register : glance | Granting/revoking user roles ------------- 3.62s\nservice-ks-register : glance | Creating users --------------------------- 3.58s\nglance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.55s\nservice-cert-copy : glance | Copying over extra CA certificates --------- 3.49s\nservice-ks-register : glance | Creating projects ------------------------ 3.46s\nservice-check-containers : Include tasks -------------------------------- 3.44s\nglance : Creating TLS backend PEM File ---------------------------------- 3.42s\nservice-ks-register : glance | Creating/deleting services --------------- 3.36s\nglance : Copying over glance-cache.conf for glance_api ------------------ 3.14s\nservice-ks-register : glance | Creating roles --------------------------- 3.14s\nservice-cert-copy : glance | Copying over backend internal TLS key ------ 3.12s\n\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:05:34 +0000 (0:00:00.332)       0:00:00.332 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:05:35 +0000 (0:00:00.472)       0:00:00.805 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m\n\nPLAY [Apply role cinder] *******************************************************\n\nTASK [cinder : include_tasks] **************************************************\nFriday 13 March 2026  01:05:35 +0000 (0:00:00.331)       0:00:01.137 ********** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : cinder | Creating/deleting services] ***************\nFriday 13 March 2026  01:05:35 +0000 (0:00:00.547)       0:00:01.684 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder (block-storage))\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m\n\nTASK [service-ks-register : cinder | Creating/deleting endpoints] **************\nFriday 13 March 2026  01:05:41 +0000 (0:00:05.941)       0:00:07.625 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> https://api-int.testbed.osism.xyz:8776/v3 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> https://api.testbed.osism.xyz:8776/v3 -> public)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating projects] ************************\nFriday 13 March 2026  01:05:54 +0000 (0:00:12.584)       0:00:20.210 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating users] ***************************\nFriday 13 March 2026  01:05:57 +0000 (0:00:03.268)       0:00:23.478 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : cinder | Creating roles] ***************************\nFriday 13 March 2026  01:06:01 +0000 (0:00:03.499)       0:00:26.977 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : cinder | Granting/revoking user roles] *************\nFriday 13 March 2026  01:06:04 +0000 (0:00:03.438)       0:00:30.416 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m\n\nTASK [cinder : Ensuring config directories exist] ******************************\nFriday 13 March 2026  01:06:11 +0000 (0:00:06.516)       0:00:36.932 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nFriday 13 March 2026  01:06:13 +0000 (0:00:02.596)       0:00:39.529 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nFriday 13 March 2026  01:06:14 +0000 (0:00:00.354)       0:00:39.883 ********** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [cinder : Ensuring cinder service ceph config subdirs exists] *************\nFriday 13 March 2026  01:06:14 +0000 (0:00:00.762)       0:00:40.645 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Copying over multiple ceph.conf for cinder services] ************\nFriday 13 March 2026  01:06:16 +0000 (0:00:01.866)       0:00:42.512 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************\nFriday 13 March 2026  01:06:22 +0000 (0:00:05.972)       0:00:48.485 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************\nFriday 13 March 2026  01:06:25 +0000 (0:00:02.868)       0:00:51.354 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m\n\nTASK [cinder : Ensuring config directory has correct owner and permission] *****\nFriday 13 March 2026  01:06:29 +0000 (0:00:03.729)       0:00:55.083 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Check if policies shall be overwritten] *************************\nFriday 13 March 2026  01:06:30 +0000 (0:00:01.206)       0:00:56.290 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Set cinder policy file] *****************************************\nFriday 13 March 2026  01:06:30 +0000 (0:00:00.120)       0:00:56.410 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nFriday 13 March 2026  01:06:31 +0000 (0:00:00.331)       0:00:56.741 ********** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over extra CA certificates] *********\nFriday 13 March 2026  01:06:31 +0000 (0:00:00.713)       0:00:57.454 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:06:34 +0000 (0:00:03.162)       0:01:00.617 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******\nFriday 13 March 2026  01:06:35 +0000 (0:00:00.669)       0:01:01.286 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over config.json files for services] ********************\nFriday 13 March 2026  01:06:36 +0000 (0:00:01.233)       0:01:02.520 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Copying over cinder-wsgi.conf] **********************************\nFriday 13 March 2026  01:06:40 +0000 (0:00:03.918)       0:01:06.438 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Configure uWSGI for Cinder] **********************************************\nFriday 13 March 2026  01:06:41 +0000 (0:00:00.714)       0:01:07.153 ********** \n\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-uwsgi-config : Copying over cinder-api uWSGI config] *************\nFriday 13 March 2026  01:06:42 +0000 (0:00:01.436)       0:01:08.590 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [cinder : Copying over cinder.conf] ***************************************\nFriday 13 March 2026  01:06:45 +0000 (0:00:02.748)       0:01:11.339 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************\nFriday 13 March 2026  01:06:55 +0000 (0:00:10.013)       0:01:21.352 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over existing policy file] ******************************\nFriday 13 March 2026  01:06:57 +0000 (0:00:01.612)       0:01:22.965 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over nfs_shares files for cinder_volume] ****************\nFriday 13 March 2026  01:06:57 +0000 (0:00:00.673)       0:01:23.638 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : cinder | Check containers] ********************\nFriday 13 March 2026  01:06:58 +0000 (0:00:00.344)       0:01:23.982 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : cinder | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:07:01 +0000 (0:00:03.045)       0:01:27.028 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:07:01 +0000 (0:00:00.564)       0:01:27.592 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nFriday 13 March 2026  01:07:02 +0000 (0:00:00.875)       0:01:28.467 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Creating Cinder database] ***************************************\nFriday 13 March 2026  01:07:03 +0000 (0:00:00.293)       0:01:28.761 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Creating Cinder database user and setting permissions] **********\nFriday 13 March 2026  01:07:04 +0000 (0:00:01.804)       0:01:30.565 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Running Cinder bootstrap container] *****************************\nFriday 13 March 2026  01:07:07 +0000 (0:00:02.545)       0:01:33.111 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:25 +0000 (0:00:18.031)       0:01:51.143 ********** \n\nTASK [cinder : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:25 +0000 (0:00:00.066)       0:01:51.210 ********** \n\nTASK [cinder : Flush handlers] *************************************************\nFriday 13 March 2026  01:07:25 +0000 (0:00:00.078)       0:01:51.289 ********** \n\nRUNNING HANDLER [cinder : Restart cinder-api container] ************************\nFriday 13 March 2026  01:07:25 +0000 (0:00:00.066)       0:01:51.355 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************\nFriday 13 March 2026  01:07:49 +0000 (0:00:23.444)       0:02:14.799 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-volume container] *********************\nFriday 13 March 2026  01:07:54 +0000 (0:00:05.290)       0:02:20.089 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-backup container] *********************\nFriday 13 March 2026  01:08:21 +0000 (0:00:27.364)       0:02:47.454 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***\nFriday 13 March 2026  01:08:32 +0000 (0:00:10.790)       0:02:58.244 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:08:32 +0000 (0:00:00.295)       0:02:58.540 ********** \n=============================================================================== \ncinder : Restart cinder-volume container ------------------------------- 27.36s\ncinder : Restart cinder-api container ---------------------------------- 23.44s\ncinder : Running Cinder bootstrap container ---------------------------- 18.03s\nservice-ks-register : cinder | Creating/deleting endpoints ------------- 12.58s\ncinder : Restart cinder-backup container ------------------------------- 10.79s\ncinder : Copying over cinder.conf -------------------------------------- 10.01s\nservice-ks-register : cinder | Granting/revoking user roles ------------- 6.52s\ncinder : Copying over multiple ceph.conf for cinder services ------------ 5.97s\nservice-ks-register : cinder | Creating/deleting services --------------- 5.94s\ncinder : Restart cinder-scheduler container ----------------------------- 5.29s\ncinder : Copying over config.json files for services -------------------- 3.92s\ncinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.73s\nservice-ks-register : cinder | Creating users --------------------------- 3.50s\nservice-ks-register : cinder | Creating roles --------------------------- 3.44s\nservice-ks-register : cinder | Creating projects ------------------------ 3.27s\nservice-cert-copy : cinder | Copying over extra CA certificates --------- 3.16s\nservice-check-containers : cinder | Check containers -------------------- 3.05s\ncinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.87s\nservice-uwsgi-config : Copying over cinder-api uWSGI config ------------- 2.75s\ncinder : Ensuring config directories exist ------------------------------ 2.60s\n\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:07:35 +0000 (0:00:00.254)       0:00:00.255 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:07:35 +0000 (0:00:00.297)       0:00:00.552 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m\n\nPLAY [Apply role grafana] ******************************************************\n\nTASK [grafana : include_tasks] *************************************************\nFriday 13 March 2026  01:07:36 +0000 (0:00:00.414)       0:00:00.967 ********** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [grafana : Ensuring config directories exist] *****************************\nFriday 13 March 2026  01:07:36 +0000 (0:00:00.487)       0:00:01.454 ********** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [grafana : Check if extra configuration file exists] **********************\nFriday 13 March 2026  01:07:37 +0000 (0:00:00.665)       0:00:02.119 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : include_tasks] *************************************************\nFriday 13 March 2026  01:07:38 +0000 (0:00:00.884)       0:00:03.004 ********** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over extra CA certificates] ********\nFriday 13 March 2026  01:07:38 +0000 (0:00:00.647)       0:00:03.652 ********** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:07:40 +0000 (0:00:01.267)       0:00:04.920 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****\nFriday 13 March 2026  01:07:40 +0000 (0:00:00.490)       0:00:05.411 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over config.json files] ********************************\nFriday 13 March 2026  01:07:41 +0000 (0:00:00.874)       0:00:06.286 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [grafana : Copying over grafana.ini] **************************************\nFriday 13 March 2026  01:07:42 +0000 (0:00:01.145)       0:00:07.431 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [grafana : Copying over extra configuration file] *************************\nFriday 13 March 2026  01:07:44 +0000 (0:00:01.230)       0:00:08.662 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Configuring Prometheus as data source for Grafana] *************\nFriday 13 March 2026  01:07:44 +0000 (0:00:00.458)       0:00:09.120 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\nTASK [grafana : Configuring dashboards provisioning] ***************************\nFriday 13 March 2026  01:07:45 +0000 (0:00:01.181)       0:00:10.302 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\nTASK [grafana : Check if the folder for custom grafana dashboards exists] ******\nFriday 13 March 2026  01:07:46 +0000 (0:00:01.104)       0:00:11.406 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : Remove templated Grafana dashboards] ***************************\nFriday 13 March 2026  01:07:47 +0000 (0:00:00.746)       0:00:12.152 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over custom dashboards] ********************************\nFriday 13 March 2026  01:07:48 +0000 (0:00:00.692)       0:00:12.845 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : grafana | Check containers] *******************\nFriday 13 March 2026  01:07:49 +0000 (0:00:01.458)       0:00:14.303 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\nTASK [service-check-containers : grafana | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:07:50 +0000 (0:00:01.342)       0:00:15.646 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:07:51 +0000 (0:00:00.363)       0:00:16.009 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Creating grafana database] *************************************\nFriday 13 March 2026  01:07:52 +0000 (0:00:00.709)       0:00:16.719 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Creating grafana database user and setting permissions] ********\nFriday 13 March 2026  01:07:54 +0000 (0:00:02.253)       0:00:18.973 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Flush handlers] ************************************************\nFriday 13 March 2026  01:07:56 +0000 (0:00:02.344)       0:00:21.317 ********** \n\nTASK [grafana : Flush handlers] ************************************************\nFriday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.381 ********** \n\nTASK [grafana : Flush handlers] ************************************************\nFriday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.444 ********** \n\nRUNNING HANDLER [grafana : Restart first grafana container] ********************\nFriday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.507 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********\nFriday 13 March 2026  01:08:03 +0000 (0:00:07.102)       0:00:28.610 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Restart remaining grafana containers] ***************\nFriday 13 March 2026  01:08:17 +0000 (0:00:13.893)       0:00:42.503 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Wait for grafana application ready] ****************************\nFriday 13 March 2026  01:08:52 +0000 (0:00:34.933)       0:01:17.437 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Remove old grafana docker volume] ******************************\nFriday 13 March 2026  01:08:54 +0000 (0:00:01.878)       0:01:19.316 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Enable grafana datasources] ************************************\nFriday 13 March 2026  01:08:54 +0000 (0:00:00.279)       0:01:19.595 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m\n\nTASK [grafana : Disable Getting Started panel] *********************************\nFriday 13 March 2026  01:08:56 +0000 (0:00:01.982)       0:01:21.577 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:08:57 +0000 (0:00:00.250)       0:01:21.828 ********** \n=============================================================================== \ngrafana : Restart remaining grafana containers ------------------------- 34.93s\ngrafana : Waiting for grafana to start on first node ------------------- 13.89s\ngrafana : Restart first grafana container ------------------------------- 7.10s\ngrafana : Creating grafana database user and setting permissions -------- 2.34s\ngrafana : Creating grafana database ------------------------------------- 2.25s\ngrafana : Enable grafana datasources ------------------------------------ 1.98s\ngrafana : Wait for grafana application ready ---------------------------- 1.88s\ngrafana : Copying over custom dashboards -------------------------------- 1.46s\nservice-check-containers : grafana | Check containers ------------------- 1.34s\nservice-cert-copy : grafana | Copying over extra CA certificates -------- 1.27s\ngrafana : Copying over grafana.ini -------------------------------------- 1.23s\ngrafana : Configuring Prometheus as data source for Grafana ------------- 1.18s\ngrafana : Copying over config.json files -------------------------------- 1.15s\ngrafana : Configuring dashboards provisioning --------------------------- 1.10s\ngrafana : Check if extra configuration file exists ---------------------- 0.88s\nservice-cert-copy : grafana | Copying over backend internal TLS key ----- 0.87s\ngrafana : Check if the folder for custom grafana dashboards exists ------ 0.75s\nservice-check-containers : Include tasks -------------------------------- 0.71s\ngrafana : Remove templated Grafana dashboards --------------------------- 0.69s\ngrafana : Ensuring config directories exist ----------------------------- 0.67s\n\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m\n\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:07:12 +0000 (0:00:00.130)       0:00:00.130 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:07:13 +0000 (0:00:00.221)       0:00:00.352 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Wait for the Nova service] ***********************************************\n\nTASK [Waiting for Nova public port to be UP] ***********************************\nFriday 13 March 2026  01:07:13 +0000 (0:00:00.554)       0:00:00.907 ********** \n\nSTILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:10:07 +0000 (0:02:53.817)       0:02:54.724 ********** \n=============================================================================== \nWaiting for Nova public port to be UP --------------------------------- 173.82s\nGroup hosts based on enabled services ----------------------------------- 0.55s\nGroup hosts based on Kolla action --------------------------------------- 0.22s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on OpenStack release] **********************************\nFriday 13 March 2026  01:05:59 +0000 (0:00:00.612)       0:00:00.612 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:06:01 +0000 (0:00:01.949)       0:00:02.561 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:06:02 +0000 (0:00:01.465)       0:00:04.027 ********** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Bootstrap nova API databases] ********************************************\n\nTASK [Bootstrap deploy] ********************************************************\nFriday 13 March 2026  01:06:03 +0000 (0:00:00.706)       0:00:04.734 ********** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Creating Nova databases] ******************************************\nFriday 13 March 2026  01:06:03 +0000 (0:00:00.549)       0:00:05.283 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m\n\nTASK [nova : Creating Nova databases user and setting permissions] *************\nFriday 13 March 2026  01:06:07 +0000 (0:00:03.931)       0:00:09.215 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nFriday 13 March 2026  01:06:11 +0000 (0:00:03.467)       0:00:12.682 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over config.json files for nova-api-bootstrap] ************\nFriday 13 March 2026  01:06:11 +0000 (0:00:00.746)       0:00:13.429 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************\nFriday 13 March 2026  01:06:13 +0000 (0:00:01.675)       0:00:15.104 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nFriday 13 March 2026  01:06:16 +0000 (0:00:03.180)       0:00:18.285 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nFriday 13 March 2026  01:06:17 +0000 (0:00:00.430)       0:00:18.715 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Create cell0 mappings] ********************************************\nFriday 13 March 2026  01:06:46 +0000 (0:00:28.997)       0:00:47.712 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nFriday 13 March 2026  01:07:01 +0000 (0:00:14.834)       0:01:02.547 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nFriday 13 March 2026  01:07:14 +0000 (0:00:13.131)       0:01:15.679 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Update cell0 mappings] ********************************************\nFriday 13 March 2026  01:07:15 +0000 (0:00:00.904)       0:01:16.583 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nFriday 13 March 2026  01:07:15 +0000 (0:00:00.456)       0:01:17.040 ********** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nFriday 13 March 2026  01:07:15 +0000 (0:00:00.393)       0:01:17.433 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nFriday 13 March 2026  01:07:33 +0000 (0:00:17.826)       0:01:35.260 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Bootstrap nova cell databases] *******************************************\n\nTASK [Bootstrap deploy] ********************************************************\nFriday 13 March 2026  01:07:34 +0000 (0:00:00.310)       0:01:35.570 ********** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database] *********************************\nFriday 13 March 2026  01:07:34 +0000 (0:00:00.560)       0:01:36.130 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database user and setting permissions] ****\nFriday 13 March 2026  01:07:36 +0000 (0:00:01.929)       0:01:38.060 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nFriday 13 March 2026  01:07:38 +0000 (0:00:01.833)       0:01:39.893 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nFriday 13 March 2026  01:07:38 +0000 (0:00:00.333)       0:01:40.226 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nFriday 13 March 2026  01:07:48 +0000 (0:00:09.884)       0:01:50.111 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nFriday 13 March 2026  01:07:48 +0000 (0:00:00.331)       0:01:50.442 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nFriday 13 March 2026  01:07:49 +0000 (0:00:00.731)       0:01:51.174 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******\nFriday 13 March 2026  01:07:50 +0000 (0:00:00.886)       0:01:52.060 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************\nFriday 13 March 2026  01:07:51 +0000 (0:00:00.917)       0:01:52.978 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Running Nova cell bootstrap container] ***********************\nFriday 13 March 2026  01:07:53 +0000 (0:00:01.969)       0:01:54.947 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nFriday 13 March 2026  01:08:14 +0000 (0:00:21.106)       0:02:16.053 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nFriday 13 March 2026  01:08:25 +0000 (0:00:11.152)       0:02:27.206 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Create cell] *************************************************\nFriday 13 March 2026  01:08:26 +0000 (0:00:00.898)       0:02:28.104 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Update cell] *************************************************\nFriday 13 March 2026  01:08:38 +0000 (0:00:11.932)       0:02:40.036 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nFriday 13 March 2026  01:08:39 +0000 (0:00:01.016)       0:02:41.053 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova] *********************************************************\n\nTASK [nova : include_tasks] ****************************************************\nFriday 13 March 2026  01:08:40 +0000 (0:00:00.508)       0:02:41.562 ********** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : nova | Creating/deleting services] *****************\nFriday 13 March 2026  01:08:40 +0000 (0:00:00.531)       0:02:42.093 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m\n\nTASK [service-ks-register : nova | Creating/deleting endpoints] ****************\nFriday 13 March 2026  01:08:43 +0000 (0:00:03.067)       0:02:45.161 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m\n\nTASK [service-ks-register : nova | Creating projects] **************************\nFriday 13 March 2026  01:08:49 +0000 (0:00:05.679)       0:02:50.840 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating users] *****************************\nFriday 13 March 2026  01:08:52 +0000 (0:00:02.717)       0:02:53.558 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : nova | Creating roles] *****************************\nFriday 13 March 2026  01:08:55 +0000 (0:00:03.392)       0:02:56.950 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : nova | Granting/revoking user roles] ***************\nFriday 13 March 2026  01:08:58 +0000 (0:00:02.769)       0:02:59.719 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nFriday 13 March 2026  01:09:06 +0000 (0:00:08.218)       0:03:07.937 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Check if policies shall be overwritten] ***************************\nFriday 13 March 2026  01:09:08 +0000 (0:00:01.715)       0:03:09.653 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : Set nova policy file] *********************************************\nFriday 13 March 2026  01:09:08 +0000 (0:00:00.143)       0:03:09.797 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check for vendordata file] ****************************************\nFriday 13 March 2026  01:09:08 +0000 (0:00:00.470)       0:03:10.267 ********** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [nova : Set vendordata file path] *****************************************\nFriday 13 March 2026  01:09:09 +0000 (0:00:00.794)       0:03:11.062 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nFriday 13 March 2026  01:09:09 +0000 (0:00:00.293)       0:03:11.356 ********** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nFriday 13 March 2026  01:09:10 +0000 (0:00:00.547)       0:03:11.903 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:09:13 +0000 (0:00:03.231)       0:03:15.135 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nFriday 13 March 2026  01:09:14 +0000 (0:00:00.824)       0:03:15.959 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over config.json files for services] **********************\nFriday 13 March 2026  01:09:15 +0000 (0:00:00.916)       0:03:16.876 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over nova.conf] *******************************************\nFriday 13 March 2026  01:09:19 +0000 (0:00:03.913)       0:03:20.789 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over existing policy file] ********************************\nFriday 13 March 2026  01:09:26 +0000 (0:00:07.532)       0:03:28.321 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-api-wsgi.conf] **********************************\nFriday 13 March 2026  01:09:27 +0000 (0:00:00.714)       0:03:29.036 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-metadata-wsgi.conf] *****************************\nFriday 13 March 2026  01:09:28 +0000 (0:00:00.648)       0:03:29.685 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over vendordata file for nova services] *******************\nFriday 13 March 2026  01:09:29 +0000 (0:00:01.008)       0:03:30.694 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-metadata) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-metadata) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-metadata) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Configure uWSGI for Nova] ************************************************\nFriday 13 March 2026  01:09:29 +0000 (0:00:00.591)       0:03:31.285 ********** \n\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova-api', 'port': '8774', 'workers': '2'})\u001b[0m\n\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova-metadata', 'port': '8775', 'workers': '2'})\u001b[0m\n\nTASK [service-uwsgi-config : Copying over nova-api uWSGI config] ***************\nFriday 13 March 2026  01:09:31 +0000 (0:00:01.372)       0:03:32.657 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [service-uwsgi-config : Copying over nova-metadata uWSGI config] **********\nFriday 13 March 2026  01:09:33 +0000 (0:00:02.373)       0:03:35.030 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [service-check-containers : nova | Check containers] **********************\nFriday 13 March 2026  01:09:35 +0000 (0:00:02.216)       0:03:37.247 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : nova | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:09:38 +0000 (0:00:02.910)       0:03:40.158 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:09:39 +0000 (0:00:00.701)       0:03:40.859 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Flush handlers] ***************************************************\nFriday 13 March 2026  01:09:40 +0000 (0:00:00.932)       0:03:41.792 ********** \n\nTASK [nova : Flush handlers] ***************************************************\nFriday 13 March 2026  01:09:40 +0000 (0:00:00.171)       0:03:41.964 ********** \n\nTASK [nova : Flush handlers] ***************************************************\nFriday 13 March 2026  01:09:40 +0000 (0:00:00.131)       0:03:42.095 ********** \n\nRUNNING HANDLER [nova : Restart nova-scheduler container] **********************\nFriday 13 March 2026  01:09:40 +0000 (0:00:00.224)       0:03:42.320 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-api container] ****************************\nFriday 13 March 2026  01:09:59 +0000 (0:00:18.797)       0:04:01.118 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-metadata container] ***********************\nFriday 13 March 2026  01:10:05 +0000 (0:00:05.886)       0:04:07.005 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova-cell] ****************************************************\n\nTASK [nova-cell : include_tasks] ***********************************************\nFriday 13 March 2026  01:10:10 +0000 (0:00:04.726)       0:04:11.731 ********** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nFriday 13 March 2026  01:10:11 +0000 (0:00:01.094)       0:04:12.825 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Get new Libvirt version] *************************************\nFriday 13 March 2026  01:10:11 +0000 (0:00:00.549)       0:04:13.374 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Cache new Libvirt version] ***********************************\nFriday 13 March 2026  01:10:30 +0000 (0:00:18.866)       0:04:32.241 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get nova_libvirt image info] *********************************************\nFriday 13 March 2026  01:10:31 +0000 (0:00:01.186)       0:04:33.427 ********** \n\u001b[0;36mincluded: service-image-info for testbed-node-3\u001b[0m\n\nTASK [service-image-info : community.docker.docker_image_info] *****************\nFriday 13 March 2026  01:10:32 +0000 (0:00:00.654)       0:04:34.082 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [service-image-info : set_fact] *******************************************\nFriday 13 March 2026  01:10:35 +0000 (0:00:02.933)       0:04:37.016 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [service-image-info : containers.podman.podman_image_info] ****************\nFriday 13 March 2026  01:10:37 +0000 (0:00:01.922)       0:04:38.938 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [service-image-info : set_fact] *******************************************\nFriday 13 March 2026  01:10:39 +0000 (0:00:02.002)       0:04:40.941 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Get container facts] *****************************************\nFriday 13 March 2026  01:10:41 +0000 (0:00:01.691)       0:04:42.633 ********** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\nTASK [nova-cell : Get current Libvirt version] *********************************\nFriday 13 March 2026  01:10:49 +0000 (0:00:08.538)       0:04:51.171 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Check that the new Libvirt version is >= current] ************\nFriday 13 March 2026  01:10:54 +0000 (0:00:05.194)       0:04:56.366 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-3', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-4', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-5', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Load and persist br_netfilter module] ************************************\nFriday 13 March 2026  01:10:58 +0000 (0:00:03.362)       0:04:59.728 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nFriday 13 March 2026  01:10:59 +0000 (0:00:00.920)       0:05:00.649 ********** \n\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nFriday 13 March 2026  01:10:59 +0000 (0:00:00.653)       0:05:01.303 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nFriday 13 March 2026  01:11:00 +0000 (0:00:01.057)       0:05:02.360 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************\nFriday 13 March 2026  01:11:01 +0000 (0:00:00.649)       0:05:03.010 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\nTASK [nova-cell : Install udev kolla kvm rules] ********************************\nFriday 13 March 2026  01:11:02 +0000 (0:00:00.999)       0:05:04.010 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Mask qemu-kvm service] ***************************************\nFriday 13 March 2026  01:11:03 +0000 (0:00:00.948)       0:05:04.958 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nFriday 13 March 2026  01:11:04 +0000 (0:00:01.462)       0:05:06.421 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nFriday 13 March 2026  01:11:06 +0000 (0:00:02.022)       0:05:08.443 ********** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nFriday 13 March 2026  01:11:08 +0000 (0:00:01.118)       0:05:09.562 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:11:11 +0000 (0:00:03.132)       0:05:12.695 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nFriday 13 March 2026  01:11:13 +0000 (0:00:01.947)       0:05:14.643 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nFriday 13 March 2026  01:11:15 +0000 (0:00:02.146)       0:05:16.789 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check nova keyring file] *************************************\nFriday 13 March 2026  01:11:16 +0000 (0:00:00.819)       0:05:17.608 ********** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Check cinder keyring file] ***********************************\nFriday 13 March 2026  01:11:17 +0000 (0:00:01.158)       0:05:18.767 ********** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Extract nova key from file] **********************************\nFriday 13 March 2026  01:11:18 +0000 (0:00:00.977)       0:05:19.744 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Extract cinder key from file] ********************************\nFriday 13 March 2026  01:11:18 +0000 (0:00:00.505)       0:05:20.250 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph nova keyring file] ****************************\nFriday 13 March 2026  01:11:19 +0000 (0:00:00.504)       0:05:20.754 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph cinder keyring file] **************************\nFriday 13 March 2026  01:11:20 +0000 (0:00:01.307)       0:05:22.062 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf] *****************************************\nFriday 13 March 2026  01:11:21 +0000 (0:00:01.115)       0:05:23.178 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m\n\nTASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************\nFriday 13 March 2026  01:11:25 +0000 (0:00:03.442)       0:05:26.620 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************\nFriday 13 March 2026  01:11:25 +0000 (0:00:00.271)       0:05:26.891 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring libvirt secrets directory exists] *******************\nFriday 13 March 2026  01:11:25 +0000 (0:00:00.400)       0:05:27.292 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Pushing nova secret xml for libvirt] *************************\nFriday 13 March 2026  01:11:26 +0000 (0:00:01.064)       0:05:28.356 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m\n\nTASK [nova-cell : Pushing secrets key for libvirt] *****************************\nFriday 13 March 2026  01:11:29 +0000 (0:00:02.783)       0:05:31.140 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************\nFriday 13 March 2026  01:11:32 +0000 (0:00:02.902)       0:05:34.042 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check qemu wrapper file] *************************************\nFriday 13 March 2026  01:11:34 +0000 (0:00:01.618)       0:05:35.661 ********** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Copy qemu wrapper] *******************************************\nFriday 13 March 2026  01:11:35 +0000 (0:00:00.912)       0:05:36.573 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Check if policies shall be overwritten] **********************\nFriday 13 March 2026  01:11:35 +0000 (0:00:00.469)       0:05:37.042 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Set nova policy file] ****************************************\nFriday 13 March 2026  01:11:35 +0000 (0:00:00.144)       0:05:37.187 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check for vendordata file] ***********************************\nFriday 13 March 2026  01:11:36 +0000 (0:00:00.592)       0:05:37.779 ********** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\nTASK [nova-cell : Set vendordata file path] ************************************\nFriday 13 March 2026  01:11:37 +0000 (0:00:00.816)       0:05:38.596 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for services] *****************\nFriday 13 March 2026  01:11:37 +0000 (0:00:00.758)       0:05:39.355 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf] **************************************\nFriday 13 March 2026  01:11:40 +0000 (0:00:03.152)       0:05:42.507 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over Nova compute provider config] *******************\nFriday 13 March 2026  01:11:47 +0000 (0:00:06.547)       0:05:49.055 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt configuration] **************************\nFriday 13 March 2026  01:11:48 +0000 (0:00:01.154)       0:05:50.209 ********** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\nTASK [nova-cell : Copying over libvirt TLS keys] *******************************\nFriday 13 March 2026  01:11:52 +0000 (0:00:03.385)       0:05:53.595 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt SASL configuration] *********************\nFriday 13 March 2026  01:11:52 +0000 (0:00:00.514)       0:05:54.110 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\nTASK [nova-cell : Copying files for nova-ssh] **********************************\nFriday 13 March 2026  01:11:57 +0000 (0:00:04.714)       0:05:58.824 ********** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [nova-cell : Copying VMware vCenter CA file] ******************************\nFriday 13 March 2026  01:12:03 +0000 (0:00:06.616)       0:06:05.440 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying 'release' file for nova_compute] *********************\nFriday 13 March 2026  01:12:04 +0000 (0:00:00.730)       0:06:06.171 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************\nFriday 13 March 2026  01:12:05 +0000 (0:00:00.571)       0:06:06.743 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Copying over existing policy file] ***************************\nFriday 13 March 2026  01:12:07 +0000 (0:00:02.278)       0:06:09.022 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Copying over vendordata file to containers] ******************\nFriday 13 March 2026  01:12:08 +0000 (0:00:01.443)       0:06:10.466 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-check-containers : nova_cell | Check containers] *****************\nFriday 13 March 2026  01:12:09 +0000 (0:00:00.771)       0:06:11.237 ********** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : nova_cell | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:12:12 +0000 (0:00:03.270)       0:06:14.508 ********** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:12:13 +0000 (0:00:00.801)       0:06:15.309 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nFriday 13 March 2026  01:12:15 +0000 (0:00:01.994)       0:06:17.304 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:16 +0000 (0:00:00.565)       0:06:17.870 ********** \n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:16 +0000 (0:00:00.125)       0:06:17.995 ********** \n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:16 +0000 (0:00:00.123)       0:06:18.119 ********** \n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:16 +0000 (0:00:00.292)       0:06:18.411 ********** \n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:17 +0000 (0:00:00.130)       0:06:18.541 ********** \n\nTASK [nova-cell : Flush handlers] **********************************************\nFriday 13 March 2026  01:12:17 +0000 (0:00:00.132)       0:06:18.674 ********** \n\nRUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************\nFriday 13 March 2026  01:12:17 +0000 (0:00:00.133)       0:06:18.807 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************\nFriday 13 March 2026  01:12:23 +0000 (0:00:06.285)       0:06:25.092 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************\nFriday 13 March 2026  01:12:35 +0000 (0:00:11.475)       0:06:36.568 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************\nFriday 13 March 2026  01:12:56 +0000 (0:00:21.859)       0:06:58.427 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************\nFriday 13 March 2026  01:13:29 +0000 (0:00:32.427)       0:07:30.855 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************\nFriday 13 March 2026  01:13:35 +0000 (0:00:06.078)       0:07:36.933 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-compute container] *******************\nFriday 13 March 2026  01:13:36 +0000 (0:00:00.663)       0:07:37.597 ********** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***\nFriday 13 March 2026  01:13:58 +0000 (0:00:22.835)       0:08:00.433 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Waiting for nova-compute services to register themselves] ****\nFriday 13 March 2026  01:13:59 +0000 (0:00:00.153)       0:08:00.587 ********** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if nova-compute service failed to register] *************\nFriday 13 March 2026  01:14:19 +0000 (0:00:20.048)       0:08:20.635 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Include discover_computes.yml] *******************************\nFriday 13 March 2026  01:14:28 +0000 (0:00:09.719)       0:08:30.355 ********** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nFriday 13 March 2026  01:14:32 +0000 (0:00:03.677)       0:08:34.032 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nFriday 13 March 2026  01:14:45 +0000 (0:00:12.980)       0:08:47.013 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if cell settings not found] *****************************\nFriday 13 March 2026  01:14:46 +0000 (0:00:01.426)       0:08:48.440 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Discover nova hosts] *****************************************\nFriday 13 March 2026  01:14:48 +0000 (0:00:01.574)       0:08:50.014 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nPLAY [Refresh nova scheduler cell cache] ***************************************\n\nTASK [nova : Refresh cell cache in nova scheduler] *****************************\nFriday 13 March 2026  01:15:01 +0000 (0:00:12.956)       0:09:02.971 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova super conductor services] *****************************\n\nTASK [nova : Reload nova super conductor services to remove RPC version pin] ***\nFriday 13 March 2026  01:15:02 +0000 (0:00:01.070)       0:09:04.042 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload Nova cell services] ***********************************************\n\nTASK [nova-cell : Reload nova cell services to remove RPC version cap] *********\nFriday 13 March 2026  01:15:03 +0000 (0:00:00.514)       0:09:04.556 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova API services] *****************************************\n\nTASK [nova : Reload nova API services to remove RPC version pin] ***************\nFriday 13 March 2026  01:15:04 +0000 (0:00:01.253)       0:09:05.810 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Run Nova API online data migrations] *************************************\n\nTASK [nova : Run Nova API online database migrations] **************************\nFriday 13 March 2026  01:15:05 +0000 (0:00:00.723)       0:09:06.534 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Run Nova cell online data migrations] ************************************\n\nTASK [nova-cell : Run Nova cell online database migrations] ********************\nFriday 13 March 2026  01:15:05 +0000 (0:00:00.807)       0:09:07.341 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=59  \u001b[0m \u001b[0;33mchanged=39  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=49  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=56  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=56  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=26  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:15:06 +0000 (0:00:00.454)       0:09:07.795 ********** \n=============================================================================== \nnova-cell : Restart nova-libvirt container ----------------------------- 32.43s\nnova : Running Nova API bootstrap container ---------------------------- 29.00s\nnova-cell : Restart nova-compute container ----------------------------- 22.84s\nnova-cell : Restart nova-ssh container --------------------------------- 21.86s\nnova-cell : Running Nova cell bootstrap container ---------------------- 21.11s\nnova-cell : Waiting for nova-compute services to register themselves --- 20.05s\nnova-cell : Get new Libvirt version ------------------------------------ 18.87s\nnova : Restart nova-scheduler container -------------------------------- 18.80s\nnova : Running Nova API bootstrap container ---------------------------- 17.83s\nnova : Create cell0 mappings ------------------------------------------- 14.83s\nnova-cell : Get a list of existing cells ------------------------------- 13.13s\nnova-cell : Get a list of existing cells ------------------------------- 12.98s\nnova-cell : Discover nova hosts ---------------------------------------- 12.96s\nnova-cell : Create cell ------------------------------------------------ 11.93s\nnova-cell : Restart nova-novncproxy container -------------------------- 11.48s\nnova-cell : Get a list of existing cells ------------------------------- 11.15s\nservice-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 9.88s\nnova-cell : Fail if nova-compute service failed to register ------------- 9.72s\nnova-cell : Get container facts ----------------------------------------- 8.54s\nservice-ks-register : nova | Granting/revoking user roles --------------- 8.22s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:10:11 +0000 (0:00:00.192)       0:00:00.192 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:10:12 +0000 (0:00:00.258)       0:00:00.451 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m\n\nPLAY [Apply role octavia] ******************************************************\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:10:12 +0000 (0:00:00.352)       0:00:00.804 ********** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : octavia | Creating/deleting services] **************\nFriday 13 March 2026  01:10:13 +0000 (0:00:00.444)       0:00:01.248 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m\n\nTASK [service-ks-register : octavia | Creating/deleting endpoints] *************\nFriday 13 March 2026  01:10:16 +0000 (0:00:03.572)       0:00:04.820 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating projects] ***********************\nFriday 13 March 2026  01:10:22 +0000 (0:00:06.287)       0:00:11.108 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating users] **************************\nFriday 13 March 2026  01:10:25 +0000 (0:00:02.773)       0:00:13.882 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : octavia | Creating roles] **************************\nFriday 13 March 2026  01:10:32 +0000 (0:00:06.910)       0:00:20.793 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : octavia | Granting/revoking user roles] ************\nFriday 13 March 2026  01:10:35 +0000 (0:00:03.083)       0:00:23.876 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\nTASK [octavia : Adding octavia related roles] **********************************\nFriday 13 March 2026  01:10:41 +0000 (0:00:06.223)       0:00:30.099 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:10:56 +0000 (0:00:14.932)       0:00:45.032 ********** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create amphora flavor] *****************************************\nFriday 13 March 2026  01:10:57 +0000 (0:00:00.538)       0:00:45.571 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create nova keypair for amphora] *******************************\nFriday 13 March 2026  01:11:02 +0000 (0:00:05.287)       0:00:50.858 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nFriday 13 March 2026  01:11:06 +0000 (0:00:03.972)       0:00:54.831 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create security groups for octavia] ****************************\nFriday 13 March 2026  01:11:09 +0000 (0:00:02.887)       0:00:57.718 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Add rules for security groups] *********************************\nFriday 13 March 2026  01:11:18 +0000 (0:00:08.871)       0:01:06.589 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m\n\nTASK [octavia : Create loadbalancer management network] ************************\nFriday 13 March 2026  01:11:32 +0000 (0:00:13.854)       0:01:20.444 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management subnet] *************************\nFriday 13 March 2026  01:11:36 +0000 (0:00:04.292)       0:01:24.737 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management router for IPv6] ****************\nFriday 13 March 2026  01:11:41 +0000 (0:00:04.631)       0:01:29.368 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update loadbalancer management subnet] *************************\nFriday 13 March 2026  01:11:41 +0000 (0:00:00.270)       0:01:29.639 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:11:45 +0000 (0:00:04.190)       0:01:33.829 ********** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m\n\nTASK [octavia : Create ports for Octavia health-manager nodes] *****************\nFriday 13 March 2026  01:11:46 +0000 (0:00:01.085)       0:01:34.914 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Update Octavia health manager port host_id] ********************\nFriday 13 March 2026  01:11:53 +0000 (0:00:06.380)       0:01:41.295 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Add Octavia port to openvswitch br-int] ************************\nFriday 13 March 2026  01:11:57 +0000 (0:00:04.125)       0:01:45.420 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Install isc-dhcp-client package] *******************************\nFriday 13 March 2026  01:11:58 +0000 (0:00:00.831)       0:01:46.251 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia dhclient conf] **********************************\nFriday 13 March 2026  01:12:00 +0000 (0:00:02.366)       0:01:48.618 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Create octavia-interface service] ******************************\nFriday 13 March 2026  01:12:01 +0000 (0:00:01.432)       0:01:50.050 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Restart octavia-interface.service if required] *****************\nFriday 13 March 2026  01:12:03 +0000 (0:00:01.289)       0:01:51.340 ********** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Enable and start octavia-interface.service] ********************\nFriday 13 March 2026  01:12:04 +0000 (0:00:01.840)       0:01:53.180 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Wait for interface ohm0 ip appear] *****************************\nFriday 13 March 2026  01:12:06 +0000 (0:00:01.930)       0:01:55.111 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Gather facts] **************************************************\nFriday 13 March 2026  01:12:07 +0000 (0:00:00.625)       0:01:55.736 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:12:10 +0000 (0:00:02.937)       0:01:58.674 ********** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Get amphora flavor info] ***************************************\nFriday 13 March 2026  01:12:11 +0000 (0:00:00.853)       0:01:59.528 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nFriday 13 March 2026  01:12:15 +0000 (0:00:04.249)       0:02:03.778 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get security groups for octavia] *******************************\nFriday 13 March 2026  01:12:18 +0000 (0:00:03.163)       0:02:06.941 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Get loadbalancer management network] ***************************\nFriday 13 March 2026  01:12:25 +0000 (0:00:06.562)       0:02:13.504 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia resources facts] ***********************************\nFriday 13 March 2026  01:12:28 +0000 (0:00:03.531)       0:02:17.036 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Ensuring config directories exist] *****************************\nFriday 13 March 2026  01:12:29 +0000 (0:00:00.309)       0:02:17.346 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Check if policies shall be overwritten] ************************\nFriday 13 March 2026  01:12:31 +0000 (0:00:02.367)       0:02:19.713 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia policy file] ***************************************\nFriday 13 March 2026  01:12:31 +0000 (0:00:00.122)       0:02:19.836 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over existing policy file] *****************************\nFriday 13 March 2026  01:12:32 +0000 (0:00:00.454)       0:02:20.291 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:12:32 +0000 (0:00:00.665)       0:02:20.957 ********** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over extra CA certificates] ********\nFriday 13 March 2026  01:12:33 +0000 (0:00:00.544)       0:02:21.501 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***\nFriday 13 March 2026  01:12:38 +0000 (0:00:05.122)       0:02:26.624 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****\nFriday 13 March 2026  01:12:39 +0000 (0:00:00.936)       0:02:27.560 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over config.json files for services] *******************\nFriday 13 March 2026  01:12:40 +0000 (0:00:00.876)       0:02:28.436 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over octavia-wsgi.conf] ********************************\nFriday 13 March 2026  01:12:45 +0000 (0:00:04.830)       0:02:33.267 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\nTASK [octavia : Copying over octavia.conf] *************************************\nFriday 13 March 2026  01:12:46 +0000 (0:00:01.779)       0:02:35.046 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over Octavia SSH key] **********************************\nFriday 13 March 2026  01:13:03 +0000 (0:00:16.588)       0:02:51.635 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-worker] ******************\nFriday 13 March 2026  01:13:04 +0000 (0:00:01.303)       0:02:52.938 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-housekeeping] ************\nFriday 13 March 2026  01:13:09 +0000 (0:00:04.480)       0:02:57.419 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-health-manager] **********\nFriday 13 March 2026  01:13:14 +0000 (0:00:05.091)       0:03:02.511 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [service-check-containers : octavia | Check containers] *******************\nFriday 13 March 2026  01:13:18 +0000 (0:00:04.624)       0:03:07.136 ********** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-check-containers : octavia | Notify handlers to restart containers] ***\nFriday 13 March 2026  01:13:22 +0000 (0:00:03.663)       0:03:10.800 ********** \n\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m\n\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nTASK [service-check-containers : Include tasks] ********************************\nFriday 13 March 2026  01:13:22 +0000 (0:00:00.346)       0:03:11.146 ********** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nFriday 13 March 2026  01:13:24 +0000 (0:00:01.290)       0:03:12.436 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Creating Octavia database] *************************************\nFriday 13 March 2026  01:13:24 +0000 (0:00:00.301)       0:03:12.738 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database] *************************\nFriday 13 March 2026  01:13:26 +0000 (0:00:02.176)       0:03:14.914 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia database user and setting permissions] ********\nFriday 13 March 2026  01:13:28 +0000 (0:00:02.108)       0:03:17.022 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database user and setting permissions] ***\nFriday 13 March 2026  01:13:30 +0000 (0:00:02.056)       0:03:19.079 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Running Octavia bootstrap container] ***************************\nFriday 13 March 2026  01:13:32 +0000 (0:00:01.841)       0:03:20.920 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Flush handlers] ************************************************\nFriday 13 March 2026  01:13:54 +0000 (0:00:21.553)       0:03:42.474 ********** \n\nTASK [octavia : Flush handlers] ************************************************\nFriday 13 March 2026  01:13:54 +0000 (0:00:00.069)       0:03:42.543 ********** \n\nTASK [octavia : Flush handlers] ************************************************\nFriday 13 March 2026  01:13:54 +0000 (0:00:00.068)       0:03:42.611 ********** \n\nRUNNING HANDLER [octavia : Restart octavia-api container] **********************\nFriday 13 March 2026  01:13:54 +0000 (0:00:00.265)       0:03:42.877 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************\nFriday 13 March 2026  01:14:08 +0000 (0:00:14.044)       0:03:56.922 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********\nFriday 13 March 2026  01:14:19 +0000 (0:00:10.911)       0:04:07.834 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************\nFriday 13 March 2026  01:14:31 +0000 (0:00:11.756)       0:04:19.591 ********** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-worker container] *******************\nFriday 13 March 2026  01:14:42 +0000 (0:00:10.783)       0:04:30.374 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=58  \u001b[0m \u001b[0;33mchanged=39  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:14:47 +0000 (0:00:05.088)       0:04:35.462 ********** \n=============================================================================== \noctavia : Running Octavia bootstrap container -------------------------- 21.55s\noctavia : Copying over octavia.conf ------------------------------------ 16.59s\noctavia : Adding octavia related roles --------------------------------- 14.93s\noctavia : Restart octavia-api container -------------------------------- 14.04s\noctavia : Add rules for security groups -------------------------------- 13.85s\noctavia : Restart octavia-health-manager container --------------------- 11.76s\noctavia : Restart octavia-driver-agent container ----------------------- 10.91s\noctavia : Restart octavia-housekeeping container ----------------------- 10.78s\noctavia : Create security groups for octavia ---------------------------- 8.87s\nservice-ks-register : octavia | Creating users -------------------------- 6.91s\noctavia : Get security groups for octavia ------------------------------- 6.56s\noctavia : Create ports for Octavia health-manager nodes ----------------- 6.38s\nservice-ks-register : octavia | Creating/deleting endpoints ------------- 6.29s\nservice-ks-register : octavia | Granting/revoking user roles ------------ 6.22s\noctavia : Create amphora flavor ----------------------------------------- 5.29s\nservice-cert-copy : octavia | Copying over extra CA certificates -------- 5.12s\noctavia : Copying certificate files for octavia-housekeeping ------------ 5.09s\noctavia : Restart octavia-worker container ------------------------------ 5.09s\noctavia : Copying over config.json files for services ------------------- 4.83s\noctavia : Create loadbalancer management subnet ------------------------- 4.63s\n\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state SUCCESS\u001b[0m\n\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-03-13 01:16:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\n\n--> DEPLOY IN A NUTSHELL -- END -- Fri Mar 13 01:16:26 UTC 2026",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "",
                                "# PULL IMAGES",
                                "",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "+ echo",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-03-13 00:42:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77abe84d-01e8-47ab-a1c1-42460e02917d (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77abe84d-01e8-47ab-a1c1-42460e02917d is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask aff5573c-3d9a-4d83-bd2d-d50ad7552953 (wipe-partitions) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:42:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task aff5573c-3d9a-4d83-bd2d-d50ad7552953 (wipe-partitions) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wipe partitions] *********************************************************",
                                "",
                                "TASK [Find all logical devices owned by UID 167] *******************************",
                                "Friday 13 March 2026  00:43:00 +0000 (0:00:00.095)       0:00:00.095 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all rook related logical devices] *********************************",
                                "Friday 13 March 2026  00:43:00 +0000 (0:00:00.534)       0:00:00.629 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Find all logical devices with prefix ceph] *******************************",
                                "Friday 13 March 2026  00:43:01 +0000 (0:00:00.369)       0:00:00.999 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all ceph related logical devices] *********************************",
                                "Friday 13 March 2026  00:43:01 +0000 (0:00:00.684)       0:00:01.684 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check device availability] ***********************************************",
                                "Friday 13 March 2026  00:43:02 +0000 (0:00:00.261)       0:00:01.945 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Wipe partitions with wipefs] *********************************************",
                                "Friday 13 March 2026  00:43:03 +0000 (0:00:01.230)       0:00:03.176 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Overwrite first 32M with zeros] ******************************************",
                                "Friday 13 March 2026  00:43:04 +0000 (0:00:01.570)       0:00:04.746 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Reload udev rules] *******************************************************",
                                "Friday 13 March 2026  00:43:07 +0000 (0:00:02.237)       0:00:06.984 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Request device events from the kernel] ***********************************",
                                "Friday 13 March 2026  00:43:07 +0000 (0:00:00.554)       0:00:07.538 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:43:08 +0000 (0:00:00.648)       0:00:08.186 ********** ",
                                "=============================================================================== ",
                                "Overwrite first 32M with zeros ------------------------------------------ 2.24s",
                                "Wipe partitions with wipefs --------------------------------------------- 1.57s",
                                "Check device availability ----------------------------------------------- 1.23s",
                                "Find all logical devices with prefix ceph ------------------------------- 0.68s",
                                "Request device events from the kernel ----------------------------------- 0.65s",
                                "Reload udev rules ------------------------------------------------------- 0.55s",
                                "Find all logical devices owned by UID 167 ------------------------------- 0.53s",
                                "Remove all rook related logical devices --------------------------------- 0.37s",
                                "Remove all ceph related logical devices --------------------------------- 0.26s",
                                "\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cc658442-32df-4315-b988-2f6b632e5043 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:43:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cc658442-32df-4315-b988-2f6b632e5043 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Friday 13 March 2026  00:43:24 +0000 (0:00:00.217)       0:00:00.217 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Friday 13 March 2026  00:43:25 +0000 (0:00:00.906)       0:00:01.123 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:43:26 +0000 (0:00:01.054)       0:00:02.177 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Friday 13 March 2026  00:43:32 +0000 (0:00:05.948)       0:00:08.126 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:43:32 +0000 (0:00:00.434)       0:00:08.561 ********** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.95s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.05s",
                                "osism.commons.facts : Create custom facts directory --------------------- 0.91s",
                                "Gather facts for all hosts ---------------------------------------------- 0.43s",
                                "\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m",
                                "\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8129fd5d-59bf-429f-8120-1b55b222bbf7 (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:43:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8129fd5d-59bf-429f-8120-1b55b222bbf7 (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:43:38 +0000 (0:00:00.246)       0:00:00.246 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:43:38 +0000 (0:00:00.213)       0:00:00.460 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:38 +0000 (0:00:00.204)       0:00:00.665 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:39 +0000 (0:00:00.453)       0:00:01.118 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:39 +0000 (0:00:00.170)       0:00:01.289 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:39 +0000 (0:00:00.161)       0:00:01.451 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:39 +0000 (0:00:00.199)       0:00:01.650 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:39 +0000 (0:00:00.178)       0:00:01.829 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:40 +0000 (0:00:00.174)       0:00:02.003 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:40 +0000 (0:00:00.167)       0:00:02.170 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:40 +0000 (0:00:00.186)       0:00:02.357 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:40 +0000 (0:00:00.182)       0:00:02.540 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:40 +0000 (0:00:00.353)       0:00:02.894 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:41 +0000 (0:00:00.501)       0:00:03.395 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:42 +0000 (0:00:00.536)       0:00:03.931 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:42 +0000 (0:00:00.704)       0:00:04.636 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:43 +0000 (0:00:00.295)       0:00:04.931 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:43 +0000 (0:00:00.336)       0:00:05.267 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:43 +0000 (0:00:00.182)       0:00:05.450 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:43 +0000 (0:00:00.183)       0:00:05.633 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:43 +0000 (0:00:00.177)       0:00:05.811 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:44 +0000 (0:00:00.169)       0:00:05.980 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:44 +0000 (0:00:00.185)       0:00:06.165 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:44 +0000 (0:00:00.180)       0:00:06.346 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:44 +0000 (0:00:00.172)       0:00:06.518 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:44 +0000 (0:00:00.186)       0:00:06.704 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:45 +0000 (0:00:00.777)       0:00:07.483 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:45 +0000 (0:00:00.170)       0:00:07.653 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:45 +0000 (0:00:00.172)       0:00:07.826 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.169)       0:00:07.995 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.193)       0:00:08.189 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.149)       0:00:08.339 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.109)       0:00:08.448 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.115)       0:00:08.563 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.115)       0:00:08.679 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Friday 13 March 2026  00:43:46 +0000 (0:00:00.123)       0:00:08.802 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.150)       0:00:08.953 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.139)       0:00:09.092 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.272)       0:00:09.365 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.161)       0:00:09.526 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.116)       0:00:09.643 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.115)       0:00:09.758 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Friday 13 March 2026  00:43:47 +0000 (0:00:00.110)       0:00:09.869 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.131)       0:00:10.001 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.123)       0:00:10.124 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.129)       0:00:10.254 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.123)       0:00:10.377 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.122)       0:00:10.500 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Friday 13 March 2026  00:43:48 +0000 (0:00:00.126)       0:00:10.626 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-34f0c902-d036-5088-bfac-53928dad7c62\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Friday 13 March 2026  00:43:49 +0000 (0:00:00.327)       0:00:10.953 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:43:50 +0000 (0:00:01.604)       0:00:12.558 ********** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:43:50 +0000 (0:00:00.232)       0:00:12.791 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:51 +0000 (0:00:00.210)       0:00:13.001 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:51 +0000 (0:00:00.365)       0:00:13.367 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:51 +0000 (0:00:00.192)       0:00:13.559 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:51 +0000 (0:00:00.179)       0:00:13.739 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:51 +0000 (0:00:00.174)       0:00:13.913 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:52 +0000 (0:00:00.183)       0:00:14.097 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:52 +0000 (0:00:00.472)       0:00:14.569 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:52 +0000 (0:00:00.188)       0:00:14.757 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:53 +0000 (0:00:00.176)       0:00:14.934 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:53 +0000 (0:00:00.193)       0:00:15.127 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:53 +0000 (0:00:00.379)       0:00:15.507 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:53 +0000 (0:00:00.384)       0:00:15.891 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:54 +0000 (0:00:00.411)       0:00:16.303 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:43:54 +0000 (0:00:00.376)       0:00:16.679 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:55 +0000 (0:00:00.299)       0:00:16.979 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:55 +0000 (0:00:00.346)       0:00:17.325 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:55 +0000 (0:00:00.488)       0:00:17.814 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.179)       0:00:17.993 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.184)       0:00:18.178 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.178)       0:00:18.357 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.188)       0:00:18.545 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.179)       0:00:18.724 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:56 +0000 (0:00:00.175)       0:00:18.900 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:57 +0000 (0:00:00.183)       0:00:19.083 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:57 +0000 (0:00:00.715)       0:00:19.798 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:58 +0000 (0:00:00.196)       0:00:19.995 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:58 +0000 (0:00:00.173)       0:00:20.168 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:43:58 +0000 (0:00:00.187)       0:00:20.355 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Friday 13 March 2026  00:43:58 +0000 (0:00:00.540)       0:00:20.896 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.186)       0:00:21.082 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.112)       0:00:21.195 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.116)       0:00:21.312 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.116)       0:00:21.428 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.114)       0:00:21.543 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.143)       0:00:21.686 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Friday 13 March 2026  00:43:59 +0000 (0:00:00.134)       0:00:21.820 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.138)       0:00:21.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.123)       0:00:22.082 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.107)       0:00:22.189 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.112)       0:00:22.301 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.253)       0:00:22.555 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.134)       0:00:22.689 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Friday 13 March 2026  00:44:00 +0000 (0:00:00.152)       0:00:22.842 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Friday 13 March 2026  00:44:01 +0000 (0:00:00.259)       0:00:23.101 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Friday 13 March 2026  00:44:01 +0000 (0:00:00.099)       0:00:23.201 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Friday 13 March 2026  00:44:01 +0000 (0:00:00.120)       0:00:23.322 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Friday 13 March 2026  00:44:01 +0000 (0:00:00.125)       0:00:23.447 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-14d46521-cd63-5c40-b949-080247a26419\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-27536765-3469-5fa1-a30d-2781c44211ac\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Friday 13 March 2026  00:44:01 +0000 (0:00:00.234)       0:00:23.682 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:44:02 +0000 (0:00:01.076)       0:00:24.758 ********** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:44:03 +0000 (0:00:00.558)       0:00:25.317 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:03 +0000 (0:00:00.284)       0:00:25.601 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.346)       0:00:25.947 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.165)       0:00:26.113 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.184)       0:00:26.297 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.166)       0:00:26.464 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.167)       0:00:26.631 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:04 +0000 (0:00:00.178)       0:00:26.810 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:05 +0000 (0:00:00.180)       0:00:26.990 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:05 +0000 (0:00:00.153)       0:00:27.144 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:05 +0000 (0:00:00.141)       0:00:27.285 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:05 +0000 (0:00:00.607)       0:00:27.893 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:06 +0000 (0:00:00.358)       0:00:28.251 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:06 +0000 (0:00:00.339)       0:00:28.591 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:44:07 +0000 (0:00:00.337)       0:00:28.928 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:07 +0000 (0:00:00.255)       0:00:29.184 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:07 +0000 (0:00:00.311)       0:00:29.496 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:07 +0000 (0:00:00.186)       0:00:29.683 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:07 +0000 (0:00:00.177)       0:00:29.860 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:08 +0000 (0:00:00.207)       0:00:30.067 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:08 +0000 (0:00:00.164)       0:00:30.231 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:08 +0000 (0:00:00.179)       0:00:30.411 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:08 +0000 (0:00:00.486)       0:00:30.897 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:09 +0000 (0:00:00.198)       0:00:31.096 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:09 +0000 (0:00:00.192)       0:00:31.288 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.664)       0:00:31.952 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.207)       0:00:32.160 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.203)       0:00:32.364 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.189)       0:00:32.553 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.192)       0:00:32.746 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Friday 13 March 2026  00:44:10 +0000 (0:00:00.161)       0:00:32.907 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Friday 13 March 2026  00:44:11 +0000 (0:00:00.108)       0:00:33.016 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Friday 13 March 2026  00:44:11 +0000 (0:00:00.185)       0:00:33.201 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Friday 13 March 2026  00:44:11 +0000 (0:00:00.423)       0:00:33.625 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Friday 13 March 2026  00:44:11 +0000 (0:00:00.149)       0:00:33.775 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.202)       0:00:33.977 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.181)       0:00:34.159 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.196)       0:00:34.355 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.148)       0:00:34.504 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.134)       0:00:34.638 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Friday 13 March 2026  00:44:12 +0000 (0:00:00.168)       0:00:34.806 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Friday 13 March 2026  00:44:13 +0000 (0:00:00.175)       0:00:34.982 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Friday 13 March 2026  00:44:13 +0000 (0:00:00.159)       0:00:35.141 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Friday 13 March 2026  00:44:13 +0000 (0:00:00.128)       0:00:35.269 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Friday 13 March 2026  00:44:13 +0000 (0:00:00.150)       0:00:35.419 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Friday 13 March 2026  00:44:13 +0000 (0:00:00.180)       0:00:35.600 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Friday 13 March 2026  00:44:14 +0000 (0:00:00.361)       0:00:35.961 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Friday 13 March 2026  00:44:14 +0000 (0:00:00.143)       0:00:36.105 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-a323501f-4860-533e-be9d-aa017f55891f\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-02d8354f-c2be-5171-944a-5a76777090d4\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Friday 13 March 2026  00:44:14 +0000 (0:00:00.237)       0:00:36.342 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:44:15 +0000 (0:00:00.865)       0:00:37.208 ********** ",
                                "=============================================================================== ",
                                "Write configuration file ------------------------------------------------ 3.55s",
                                "Add known links to the list of available block devices ------------------ 1.17s",
                                "Get extra vars for Ceph configuration ----------------------------------- 1.00s",
                                "Add known partitions to the list of available block devices ------------- 0.99s",
                                "Print configuration data ------------------------------------------------ 0.80s",
                                "Add known partitions to the list of available block devices ------------- 0.78s",
                                "Add known partitions to the list of available block devices ------------- 0.72s",
                                "Add known links to the list of available block devices ------------------ 0.70s",
                                "Get initial list of available block devices ----------------------------- 0.70s",
                                "Add known partitions to the list of available block devices ------------- 0.66s",
                                "Generate shared DB/WAL VG names ----------------------------------------- 0.65s",
                                "Add known links to the list of available block devices ------------------ 0.61s",
                                "Generate lvm_volumes structure (block + wal) ---------------------------- 0.61s",
                                "Print DB devices -------------------------------------------------------- 0.61s",
                                "Add known partitions to the list of available block devices ------------- 0.54s",
                                "Set DB devices config data ---------------------------------------------- 0.54s",
                                "Print ceph_osd_devices -------------------------------------------------- 0.54s",
                                "Add known links to the list of available block devices ------------------ 0.54s",
                                "Add known links to the list of available block devices ------------------ 0.50s",
                                "Set UUIDs for OSD VGs/LVs ----------------------------------------------- 0.50s",
                                "\u001b[32m2026-03-13 00:44:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5aa282cd-b8dc-4f2c-8a92-62ef961101c0 (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-03-13 00:44:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-03-13 00:45:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "[master f8e2d0f] 2026-03-13-00-45",
                                " 1 file changed, 30 insertions(+), 9 deletions(-)",
                                "\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m",
                                "\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b90afbba-b381-46bc-9cad-41547468fc50 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:45:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b90afbba-b381-46bc-9cad-41547468fc50 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:45:10 +0000 (0:00:00.325)       0:00:00.325 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:45:10 +0000 (0:00:00.288)       0:00:00.614 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:11 +0000 (0:00:00.407)       0:00:01.022 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:11 +0000 (0:00:00.529)       0:00:01.551 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:11 +0000 (0:00:00.183)       0:00:01.734 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:12 +0000 (0:00:00.196)       0:00:01.931 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:12 +0000 (0:00:00.201)       0:00:02.133 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:12 +0000 (0:00:00.206)       0:00:02.339 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:12 +0000 (0:00:00.194)       0:00:02.533 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:12 +0000 (0:00:00.211)       0:00:02.745 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:13 +0000 (0:00:00.209)       0:00:02.955 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:13 +0000 (0:00:00.217)       0:00:03.173 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:13 +0000 (0:00:00.454)       0:00:03.627 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:14 +0000 (0:00:00.597)       0:00:04.225 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:15 +0000 (0:00:00.770)       0:00:04.995 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:16 +0000 (0:00:01.153)       0:00:06.149 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:16 +0000 (0:00:00.336)       0:00:06.485 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:16 +0000 (0:00:00.442)       0:00:06.928 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:17 +0000 (0:00:00.192)       0:00:07.120 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:17 +0000 (0:00:00.215)       0:00:07.336 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:17 +0000 (0:00:00.193)       0:00:07.529 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:17 +0000 (0:00:00.170)       0:00:07.700 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:17 +0000 (0:00:00.164)       0:00:07.864 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:18 +0000 (0:00:00.177)       0:00:08.042 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:18 +0000 (0:00:00.175)       0:00:08.217 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:18 +0000 (0:00:00.168)       0:00:08.386 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:19 +0000 (0:00:00.966)       0:00:09.353 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:19 +0000 (0:00:00.151)       0:00:09.504 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:19 +0000 (0:00:00.187)       0:00:09.692 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:19 +0000 (0:00:00.187)       0:00:09.879 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Friday 13 March 2026  00:45:20 +0000 (0:00:00.139)       0:00:10.019 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Friday 13 March 2026  00:45:20 +0000 (0:00:00.098)       0:00:10.117 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '34f0c902-d036-5088-bfac-53928dad7c62'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '74a7d81d-0a8f-5a99-882a-be09baf5f158'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Friday 13 March 2026  00:45:20 +0000 (0:00:00.154)       0:00:10.272 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Friday 13 March 2026  00:45:22 +0000 (0:00:01.758)       0:00:12.031 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Friday 13 March 2026  00:45:22 +0000 (0:00:00.172)       0:00:12.204 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Friday 13 March 2026  00:45:23 +0000 (0:00:01.423)       0:00:13.627 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Friday 13 March 2026  00:45:23 +0000 (0:00:00.131)       0:00:13.758 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Friday 13 March 2026  00:45:23 +0000 (0:00:00.121)       0:00:13.880 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.267)       0:00:14.147 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.129)       0:00:14.277 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.135)       0:00:14.413 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.120)       0:00:14.534 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.132)       0:00:14.666 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.128)       0:00:14.795 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Friday 13 March 2026  00:45:24 +0000 (0:00:00.133)       0:00:14.928 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.132)       0:00:15.060 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.136)       0:00:15.197 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.132)       0:00:15.329 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.119)       0:00:15.449 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.123)       0:00:15.572 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Friday 13 March 2026  00:45:25 +0000 (0:00:00.241)       0:00:15.814 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Friday 13 March 2026  00:45:26 +0000 (0:00:00.127)       0:00:15.942 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Friday 13 March 2026  00:45:26 +0000 (0:00:00.125)       0:00:16.068 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Friday 13 March 2026  00:45:26 +0000 (0:00:00.617)       0:00:16.685 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Friday 13 March 2026  00:45:27 +0000 (0:00:00.515)       0:00:17.200 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Friday 13 March 2026  00:45:27 +0000 (0:00:00.494)       0:00:17.695 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Friday 13 March 2026  00:45:27 +0000 (0:00:00.130)       0:00:17.826 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Friday 13 March 2026  00:45:27 +0000 (0:00:00.095)       0:00:17.921 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.098)       0:00:18.019 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.130)       0:00:18.150 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.124)       0:00:18.274 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.114)       0:00:18.389 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.223)       0:00:18.612 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.137)       0:00:18.750 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Friday 13 March 2026  00:45:28 +0000 (0:00:00.135)       0:00:18.885 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.109)       0:00:18.995 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.124)       0:00:19.120 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.138)       0:00:19.258 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.151)       0:00:19.409 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.151)       0:00:19.561 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.135)       0:00:19.696 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Friday 13 March 2026  00:45:29 +0000 (0:00:00.137)       0:00:19.833 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.137)       0:00:19.971 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.128)       0:00:20.099 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.292)       0:00:20.392 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.125)       0:00:20.517 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.129)       0:00:20.647 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Friday 13 March 2026  00:45:30 +0000 (0:00:00.154)       0:00:20.802 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Friday 13 March 2026  00:45:31 +0000 (0:00:00.142)       0:00:20.944 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Friday 13 March 2026  00:45:31 +0000 (0:00:00.140)       0:00:21.085 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Friday 13 March 2026  00:45:31 +0000 (0:00:00.139)       0:00:21.225 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:45:31 +0000 (0:00:00.148)       0:00:21.373 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:45:31 +0000 (0:00:00.493)       0:00:21.867 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Friday 13 March 2026  00:45:32 +0000 (0:00:00.501)       0:00:22.368 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Friday 13 March 2026  00:45:32 +0000 (0:00:00.153)       0:00:22.522 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'vg_name': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'vg_name': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Friday 13 March 2026  00:45:32 +0000 (0:00:00.171)       0:00:22.694 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Friday 13 March 2026  00:45:33 +0000 (0:00:00.306)       0:00:23.000 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Friday 13 March 2026  00:45:33 +0000 (0:00:00.143)       0:00:23.144 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Friday 13 March 2026  00:45:33 +0000 (0:00:00.154)       0:00:23.298 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-34f0c902-d036-5088-bfac-53928dad7c62\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-34f0c902-d036-5088-bfac-53928dad7c62\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:45:33 +0000 (0:00:00.293)       0:00:23.592 ********** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:45:33 +0000 (0:00:00.222)       0:00:23.814 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:34 +0000 (0:00:00.190)       0:00:24.004 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:34 +0000 (0:00:00.370)       0:00:24.375 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:34 +0000 (0:00:00.206)       0:00:24.581 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:34 +0000 (0:00:00.188)       0:00:24.769 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:35 +0000 (0:00:00.490)       0:00:25.259 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:35 +0000 (0:00:00.195)       0:00:25.455 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:35 +0000 (0:00:00.172)       0:00:25.627 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:35 +0000 (0:00:00.213)       0:00:25.841 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:36 +0000 (0:00:00.212)       0:00:26.054 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:36 +0000 (0:00:00.202)       0:00:26.256 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:36 +0000 (0:00:00.418)       0:00:26.675 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:37 +0000 (0:00:00.392)       0:00:27.068 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:37 +0000 (0:00:00.398)       0:00:27.467 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:38 +0000 (0:00:00.499)       0:00:27.966 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:38 +0000 (0:00:00.499)       0:00:28.465 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:39 +0000 (0:00:00.513)       0:00:28.979 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:39 +0000 (0:00:00.222)       0:00:29.201 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:39 +0000 (0:00:00.183)       0:00:29.384 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:39 +0000 (0:00:00.188)       0:00:29.573 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:39 +0000 (0:00:00.197)       0:00:29.771 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:40 +0000 (0:00:00.211)       0:00:29.983 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:40 +0000 (0:00:00.185)       0:00:30.168 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:40 +0000 (0:00:00.213)       0:00:30.382 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:40 +0000 (0:00:00.212)       0:00:30.594 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:41 +0000 (0:00:00.766)       0:00:31.361 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:41 +0000 (0:00:00.200)       0:00:31.561 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:42 +0000 (0:00:00.664)       0:00:32.225 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:45:42 +0000 (0:00:00.186)       0:00:32.411 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Friday 13 March 2026  00:45:42 +0000 (0:00:00.202)       0:00:32.614 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Friday 13 March 2026  00:45:42 +0000 (0:00:00.134)       0:00:32.749 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '14d46521-cd63-5c40-b949-080247a26419'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '27536765-3469-5fa1-a30d-2781c44211ac'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Friday 13 March 2026  00:45:43 +0000 (0:00:00.217)       0:00:32.967 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Friday 13 March 2026  00:45:44 +0000 (0:00:01.858)       0:00:34.826 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Friday 13 March 2026  00:45:45 +0000 (0:00:00.139)       0:00:34.966 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Friday 13 March 2026  00:45:46 +0000 (0:00:01.377)       0:00:36.343 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Friday 13 March 2026  00:45:46 +0000 (0:00:00.137)       0:00:36.481 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Friday 13 March 2026  00:45:46 +0000 (0:00:00.163)       0:00:36.645 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Friday 13 March 2026  00:45:46 +0000 (0:00:00.142)       0:00:36.787 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Friday 13 March 2026  00:45:46 +0000 (0:00:00.131)       0:00:36.918 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.265)       0:00:37.184 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.134)       0:00:37.318 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.124)       0:00:37.443 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.098)       0:00:37.542 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.134)       0:00:37.677 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Friday 13 March 2026  00:45:47 +0000 (0:00:00.135)       0:00:37.813 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.145)       0:00:37.958 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.136)       0:00:38.094 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.123)       0:00:38.218 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.123)       0:00:38.341 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.128)       0:00:38.470 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.135)       0:00:38.605 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Friday 13 March 2026  00:45:48 +0000 (0:00:00.287)       0:00:38.893 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Friday 13 March 2026  00:45:49 +0000 (0:00:00.561)       0:00:39.454 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Friday 13 March 2026  00:45:50 +0000 (0:00:00.514)       0:00:39.968 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Friday 13 March 2026  00:45:50 +0000 (0:00:00.529)       0:00:40.498 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Friday 13 March 2026  00:45:50 +0000 (0:00:00.154)       0:00:40.653 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Friday 13 March 2026  00:45:50 +0000 (0:00:00.112)       0:00:40.765 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Friday 13 March 2026  00:45:50 +0000 (0:00:00.118)       0:00:40.884 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.148)       0:00:41.033 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.128)       0:00:41.161 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.129)       0:00:41.291 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.130)       0:00:41.421 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.131)       0:00:41.553 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.248)       0:00:41.802 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Friday 13 March 2026  00:45:51 +0000 (0:00:00.125)       0:00:41.927 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.126)       0:00:42.054 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.143)       0:00:42.197 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.131)       0:00:42.329 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.137)       0:00:42.467 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.119)       0:00:42.586 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.120)       0:00:42.706 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:45:52 +0000 (0:00:00.110)       0:00:42.816 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.132)       0:00:42.948 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.144)       0:00:43.093 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.139)       0:00:43.232 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.150)       0:00:43.383 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.332)       0:00:43.715 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Friday 13 March 2026  00:45:53 +0000 (0:00:00.148)       0:00:43.864 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Friday 13 March 2026  00:45:54 +0000 (0:00:00.135)       0:00:43.999 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Friday 13 March 2026  00:45:54 +0000 (0:00:00.133)       0:00:44.133 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:45:54 +0000 (0:00:00.128)       0:00:44.261 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:45:54 +0000 (0:00:00.503)       0:00:44.764 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Friday 13 March 2026  00:45:55 +0000 (0:00:00.512)       0:00:45.277 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Friday 13 March 2026  00:45:55 +0000 (0:00:00.130)       0:00:45.407 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'vg_name': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'vg_name': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Friday 13 March 2026  00:45:55 +0000 (0:00:00.137)       0:00:45.545 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Friday 13 March 2026  00:45:55 +0000 (0:00:00.131)       0:00:45.676 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Friday 13 March 2026  00:45:55 +0000 (0:00:00.130)       0:00:45.806 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Friday 13 March 2026  00:45:56 +0000 (0:00:00.125)       0:00:45.932 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-14d46521-cd63-5c40-b949-080247a26419\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-27536765-3469-5fa1-a30d-2781c44211ac\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-14d46521-cd63-5c40-b949-080247a26419\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-27536765-3469-5fa1-a30d-2781c44211ac\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  00:45:56 +0000 (0:00:00.379)       0:00:46.311 ********** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Friday 13 March 2026  00:45:56 +0000 (0:00:00.224)       0:00:46.535 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:56 +0000 (0:00:00.207)       0:00:46.743 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.352)       0:00:47.096 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.145)       0:00:47.242 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.136)       0:00:47.378 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.156)       0:00:47.535 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.165)       0:00:47.700 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:57 +0000 (0:00:00.164)       0:00:47.865 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:58 +0000 (0:00:00.457)       0:00:48.322 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:58 +0000 (0:00:00.180)       0:00:48.503 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:58 +0000 (0:00:00.150)       0:00:48.654 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:59 +0000 (0:00:00.381)       0:00:49.036 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:59 +0000 (0:00:00.379)       0:00:49.415 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:45:59 +0000 (0:00:00.387)       0:00:49.802 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Friday 13 March 2026  00:46:00 +0000 (0:00:00.393)       0:00:50.196 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:00 +0000 (0:00:00.312)       0:00:50.508 ********** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:00 +0000 (0:00:00.386)       0:00:50.895 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:01 +0000 (0:00:00.181)       0:00:51.077 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:01 +0000 (0:00:00.509)       0:00:51.586 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:01 +0000 (0:00:00.201)       0:00:51.788 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:02 +0000 (0:00:00.174)       0:00:51.963 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:02 +0000 (0:00:00.181)       0:00:52.144 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:02 +0000 (0:00:00.183)       0:00:52.327 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:02 +0000 (0:00:00.178)       0:00:52.506 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:02 +0000 (0:00:00.175)       0:00:52.681 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:03 +0000 (0:00:00.622)       0:00:53.304 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:03 +0000 (0:00:00.204)       0:00:53.508 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:03 +0000 (0:00:00.200)       0:00:53.709 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Friday 13 March 2026  00:46:03 +0000 (0:00:00.182)       0:00:53.891 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Friday 13 March 2026  00:46:04 +0000 (0:00:00.185)       0:00:54.077 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Friday 13 March 2026  00:46:04 +0000 (0:00:00.244)       0:00:54.322 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'a323501f-4860-533e-be9d-aa017f55891f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '02d8354f-c2be-5171-944a-5a76777090d4'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Friday 13 March 2026  00:46:04 +0000 (0:00:00.188)       0:00:54.510 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Friday 13 March 2026  00:46:06 +0000 (0:00:01.806)       0:00:56.317 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Friday 13 March 2026  00:46:06 +0000 (0:00:00.146)       0:00:56.464 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Friday 13 March 2026  00:46:07 +0000 (0:00:01.359)       0:00:57.823 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.140)       0:00:57.963 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.122)       0:00:58.086 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.117)       0:00:58.204 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.126)       0:00:58.330 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.120)       0:00:58.451 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.109)       0:00:58.561 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Friday 13 March 2026  00:46:08 +0000 (0:00:00.142)       0:00:58.703 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.269)       0:00:58.972 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.163)       0:00:59.135 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.143)       0:00:59.279 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.117)       0:00:59.397 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.100)       0:00:59.497 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.111)       0:00:59.609 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.103)       0:00:59.713 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Friday 13 March 2026  00:46:09 +0000 (0:00:00.131)       0:00:59.844 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Friday 13 March 2026  00:46:10 +0000 (0:00:00.125)       0:00:59.969 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Friday 13 March 2026  00:46:10 +0000 (0:00:00.125)       0:01:00.095 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Friday 13 March 2026  00:46:10 +0000 (0:00:00.544)       0:01:00.640 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Friday 13 March 2026  00:46:11 +0000 (0:00:00.539)       0:01:01.179 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Friday 13 March 2026  00:46:11 +0000 (0:00:00.652)       0:01:01.831 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.140)       0:01:01.971 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.098)       0:01:02.070 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.104)       0:01:02.175 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.125)       0:01:02.301 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.112)       0:01:02.413 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.119)       0:01:02.533 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.112)       0:01:02.645 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.109)       0:01:02.755 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Friday 13 March 2026  00:46:12 +0000 (0:00:00.110)       0:01:02.865 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.123)       0:01:02.988 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.122)       0:01:03.110 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.260)       0:01:03.370 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.123)       0:01:03.494 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.116)       0:01:03.610 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.112)       0:01:03.722 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Friday 13 March 2026  00:46:13 +0000 (0:00:00.116)       0:01:03.838 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.128)       0:01:03.966 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.124)       0:01:04.091 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.143)       0:01:04.235 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.153)       0:01:04.388 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.138)       0:01:04.527 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.140)       0:01:04.667 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Friday 13 March 2026  00:46:14 +0000 (0:00:00.132)       0:01:04.800 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Friday 13 March 2026  00:46:15 +0000 (0:00:00.269)       0:01:05.069 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Friday 13 March 2026  00:46:15 +0000 (0:00:00.133)       0:01:05.203 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:46:15 +0000 (0:00:00.127)       0:01:05.331 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Friday 13 March 2026  00:46:15 +0000 (0:00:00.523)       0:01:05.854 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Friday 13 March 2026  00:46:16 +0000 (0:00:00.573)       0:01:06.427 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Friday 13 March 2026  00:46:16 +0000 (0:00:00.153)       0:01:06.581 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'vg_name': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'vg_name': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Friday 13 March 2026  00:46:16 +0000 (0:00:00.148)       0:01:06.729 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Friday 13 March 2026  00:46:16 +0000 (0:00:00.151)       0:01:06.881 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Friday 13 March 2026  00:46:17 +0000 (0:00:00.138)       0:01:07.020 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Friday 13 March 2026  00:46:17 +0000 (0:00:00.136)       0:01:07.156 ********** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-02d8354f-c2be-5171-944a-5a76777090d4\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-a323501f-4860-533e-be9d-aa017f55891f\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-a323501f-4860-533e-be9d-aa017f55891f\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-02d8354f-c2be-5171-944a-5a76777090d4\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:46:17 +0000 (0:00:00.159)       0:01:07.316 ********** ",
                                "=============================================================================== ",
                                "Create block VGs -------------------------------------------------------- 5.42s",
                                "Create block LVs -------------------------------------------------------- 4.16s",
                                "Gather DB VGs with total and available size in bytes -------------------- 1.72s",
                                "Gather DB+WAL VGs with total and available size in bytes ---------------- 1.68s",
                                "Get list of Ceph PVs with associated VGs -------------------------------- 1.59s",
                                "Gather WAL VGs with total and available size in bytes ------------------- 1.57s",
                                "Get list of Ceph LVs with associated VGs -------------------------------- 1.52s",
                                "Add known partitions to the list of available block devices ------------- 1.34s",
                                "Add known links to the list of available block devices ------------------ 1.25s",
                                "Add known links to the list of available block devices ------------------ 1.15s",
                                "Add known partitions to the list of available block devices ------------- 0.97s",
                                "Print LVM report data --------------------------------------------------- 0.83s",
                                "Get initial list of available block devices ----------------------------- 0.81s",
                                "Add known links to the list of available block devices ------------------ 0.77s",
                                "Add known partitions to the list of available block devices ------------- 0.77s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.74s",
                                "Add known partitions to the list of available block devices ------------- 0.66s",
                                "Print 'Create WAL LVs for ceph_wal_devices' ----------------------------- 0.63s",
                                "Add known partitions to the list of available block devices ------------- 0.62s",
                                "Add known links to the list of available block devices ------------------ 0.60s",
                                "\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9b508752-6c86-455f-88c4-f40effbcb7a8 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:46:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9b508752-6c86-455f-88c4-f40effbcb7a8 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Friday 13 March 2026  00:46:33 +0000 (0:00:00.207)       0:00:00.207 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Friday 13 March 2026  00:46:34 +0000 (0:00:00.970)       0:00:01.178 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  00:46:35 +0000 (0:00:01.089)       0:00:02.267 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Friday 13 March 2026  00:46:40 +0000 (0:00:05.209)       0:00:07.476 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:46:41 +0000 (0:00:00.448)       0:00:07.925 ********** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.21s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.09s",
                                "osism.commons.facts : Create custom facts directory --------------------- 0.97s",
                                "Gather facts for all hosts ---------------------------------------------- 0.45s",
                                "\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m",
                                "\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 715d3fd8-4a42-42a9-9503-eff1236b6ac4 (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 00:46:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 715d3fd8-4a42-42a9-9503-eff1236b6ac4 (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Friday 13 March 2026  00:46:57 +0000 (0:00:00.174)       0:00:00.174 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Friday 13 March 2026  00:46:57 +0000 (0:00:00.172)       0:00:00.347 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Friday 13 March 2026  00:46:58 +0000 (0:00:01.067)       0:00:01.414 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Friday 13 March 2026  00:47:07 +0000 (0:00:09.158)       0:00:10.573 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Friday 13 March 2026  00:47:08 +0000 (0:00:01.063)       0:00:11.636 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Friday 13 March 2026  00:47:09 +0000 (0:00:00.933)       0:00:12.570 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Write frr_config_template to temporary file] ********",
                                "Friday 13 March 2026  00:47:10 +0000 (0:00:01.180)       0:00:13.750 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***",
                                "Friday 13 March 2026  00:47:11 +0000 (0:00:00.155)       0:00:13.906 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Remove temporary frr_config_template file] **********",
                                "Friday 13 March 2026  00:47:11 +0000 (0:00:00.142)       0:00:14.049 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Friday 13 March 2026  00:47:11 +0000 (0:00:00.154)       0:00:14.204 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Friday 13 March 2026  00:47:11 +0000 (0:00:00.146)       0:00:14.350 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Friday 13 March 2026  00:47:11 +0000 (0:00:00.143)       0:00:14.494 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Friday 13 March 2026  00:47:12 +0000 (0:00:01.135)       0:00:15.630 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Friday 13 March 2026  00:47:15 +0000 (0:00:02.180)       0:00:17.811 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.frr : Restart frr service] *********************",
                                "Friday 13 March 2026  00:47:16 +0000 (0:00:01.055)       0:00:18.866 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:47:17 +0000 (0:00:01.261)       0:00:20.127 ********** ",
                                "=============================================================================== ",
                                "osism.services.frr : Install frr package -------------------------------- 9.16s",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 2.18s",
                                "osism.services.frr : Restart frr service -------------------------------- 1.26s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.18s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.14s",
                                "osism.services.frr : Pin frr package version ---------------------------- 1.07s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.06s",
                                "osism.services.frr : Manage frr service --------------------------------- 1.06s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.93s",
                                "osism.services.frr : Include distribution specific install tasks -------- 0.17s",
                                "osism.services.frr : Write frr_config_template to temporary file -------- 0.16s",
                                "osism.services.frr : Remove temporary frr_config_template file ---------- 0.15s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 0.15s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 0.14s",
                                "osism.services.frr : Render frr.conf from frr_config_template variable --- 0.14s",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- START -- Fri Mar 13 00:47:17 UTC 2026",
                                "",
                                "\u001b[32m2026-03-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "",
                                "PLAY [Apply role geerlingguy.dotfiles] *****************************************",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****",
                                "Friday 13 March 2026  00:47:42 +0000 (0:00:00.703)       0:00:00.703 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********",
                                "Friday 13 March 2026  00:47:46 +0000 (0:00:04.267)       0:00:04.971 ********** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***",
                                "Friday 13 March 2026  00:47:49 +0000 (0:00:02.783)       0:00:07.755 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.730324', 'end': '2026-03-13 00:47:47.735404', 'delta': '0:00:00.005080', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.738730', 'end': '2026-03-13 00:47:47.742517', 'delta': '0:00:00.003787', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.732844', 'end': '2026-03-13 00:47:47.737422', 'delta': '0:00:00.004578', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:47.813782', 'end': '2026-03-13 00:47:47.818180', 'delta': '0:00:00.004398', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.778395', 'end': '2026-03-13 00:47:48.788111', 'delta': '0:00:00.009716', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.095443', 'end': '2026-03-13 00:47:48.099448', 'delta': '0:00:00.004005', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-03-13 00:47:48.744434', 'end': '2026-03-13 00:47:48.749859', 'delta': '0:00:00.005425', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****",
                                "Friday 13 March 2026  00:47:52 +0000 (0:00:03.177)       0:00:10.932 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************",
                                "Friday 13 March 2026  00:47:54 +0000 (0:00:01.708)       0:00:12.641 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:47:56 +0000 (0:00:01.905)       0:00:14.547 ********** ",
                                "=============================================================================== ",
                                "geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 4.27s",
                                "geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 3.18s",
                                "geerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.78s",
                                "geerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 1.91s",
                                "geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.71s",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac9a3ae0-9ea5-46f4-939b-113e6149507f is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:47:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f5685f16-3e39-4042-819e-d92e4489043f is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1463f566-b70c-4b9a-a842-294ab59eb2ca is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:48:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7af45cad-8d37-4a29-95ab-bcfd5ae112b1 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role homer] ********************************************************",
                                "",
                                "TASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:00.372)       0:00:00.372 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create traefik external network] ******************",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:00.245)       0:00:00.617 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create required directories] **********************",
                                "Friday 13 March 2026  00:47:43 +0000 (0:00:02.172)       0:00:02.794 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy config.yml configuration file] ***************",
                                "Friday 13 March 2026  00:47:45 +0000 (0:00:01.778)       0:00:04.573 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy docker-compose.yml file] *********************",
                                "Friday 13 March 2026  00:47:49 +0000 (0:00:04.452)       0:00:09.025 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Manage homer service] *****************************",
                                "Friday 13 March 2026  00:47:52 +0000 (0:00:02.766)       0:00:11.792 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.homer : Restart homer service] *****************",
                                "Friday 13 March 2026  00:48:19 +0000 (0:00:26.460)       0:00:38.252 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:48:21 +0000 (0:00:02.572)       0:00:40.825 ********** ",
                                "=============================================================================== ",
                                "osism.services.homer : Manage homer service ---------------------------- 26.46s",
                                "osism.services.homer : Copy config.yml configuration file --------------- 4.45s",
                                "osism.services.homer : Copy docker-compose.yml file --------------------- 2.77s",
                                "osism.services.homer : Restart homer service ---------------------------- 2.57s",
                                "osism.services.homer : Create traefik external network ------------------ 2.18s",
                                "osism.services.homer : Create required directories ---------------------- 1.78s",
                                "osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.25s",
                                "",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Friday 13 March 2026  00:47:44 +0000 (0:00:01.482)       0:00:01.482 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Friday 13 March 2026  00:47:45 +0000 (0:00:00.791)       0:00:02.274 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Friday 13 March 2026  00:47:46 +0000 (0:00:01.756)       0:00:04.030 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Friday 13 March 2026  00:47:52 +0000 (0:00:05.800)       0:00:09.830 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Friday 13 March 2026  00:48:26 +0000 (0:00:33.629)       0:00:43.459 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Friday 13 March 2026  00:48:27 +0000 (0:00:00.982)       0:00:44.442 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Friday 13 March 2026  00:48:27 +0000 (0:00:00.532)       0:00:44.975 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Friday 13 March 2026  00:48:29 +0000 (0:00:01.657)       0:00:46.632 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Friday 13 March 2026  00:48:30 +0000 (0:00:00.615)       0:00:47.247 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Friday 13 March 2026  00:48:30 +0000 (0:00:00.671)       0:00:47.919 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:48:31 +0000 (0:00:00.432)       0:00:48.351 ********** ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Manage openstackclient service -------- 33.63s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 5.80s",
                                "osism.services.openstackclient : Create required directories ------------ 1.76s",
                                "osism.services.openstackclient : Restart openstackclient service -------- 1.66s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 0.98s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.79s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.67s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 0.62s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 0.53s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 0.43s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:00.532)       0:00:00.532 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m",
                                "",
                                "PLAY [Apply role netdata] ******************************************************",
                                "",
                                "TASK [osism.services.netdata : Include distribution specific install tasks] ****",
                                "Friday 13 March 2026  00:47:44 +0000 (0:00:02.999)       0:00:03.532 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Remove old architecture-dependent repository] ***",
                                "Friday 13 March 2026  00:47:46 +0000 (0:00:01.242)       0:00:04.774 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install apt-transport-https package] ************",
                                "Friday 13 March 2026  00:47:47 +0000 (0:00:01.535)       0:00:06.309 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository gpg key] *************************",
                                "Friday 13 March 2026  00:47:50 +0000 (0:00:02.886)       0:00:09.196 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository] *********************************",
                                "Friday 13 March 2026  00:47:53 +0000 (0:00:02.593)       0:00:11.790 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install package netdata] ************************",
                                "Friday 13 March 2026  00:48:03 +0000 (0:00:10.097)       0:00:21.887 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include config tasks] ***************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:37.139)       0:00:59.027 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Copy configuration files] ***********************",
                                "Friday 13 March 2026  00:48:41 +0000 (0:00:01.399)       0:01:00.426 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***",
                                "Friday 13 March 2026  00:48:46 +0000 (0:00:05.231)       0:01:05.657 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Opt out from anonymous statistics] **************",
                                "Friday 13 March 2026  00:48:48 +0000 (0:00:01.830)       0:01:07.488 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add netdata user to docker group] ***************",
                                "Friday 13 March 2026  00:48:51 +0000 (0:00:02.267)       0:01:09.756 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Manage service netdata] *************************",
                                "Friday 13 March 2026  00:48:54 +0000 (0:00:03.529)       0:01:13.285 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include host type specific tasks] ***************",
                                "Friday 13 March 2026  00:48:57 +0000 (0:00:02.470)       0:01:15.756 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********",
                                "Friday 13 March 2026  00:48:58 +0000 (0:00:01.354)       0:01:17.111 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.netdata : Restart service netdata] *************",
                                "Friday 13 March 2026  00:49:00 +0000 (0:00:01.871)       0:01:18.982 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:49:11 +0000 (0:00:11.373)       0:01:30.356 ********** ",
                                "=============================================================================== ",
                                "osism.services.netdata : Install package netdata ----------------------- 37.14s",
                                "osism.services.netdata : Restart service netdata ----------------------- 11.37s",
                                "osism.services.netdata : Add repository -------------------------------- 10.10s",
                                "osism.services.netdata : Copy configuration files ----------------------- 5.23s",
                                "osism.services.netdata : Add netdata user to docker group --------------- 3.53s",
                                "Group hosts based on enabled services ----------------------------------- 3.00s",
                                "osism.services.netdata : Install apt-transport-https package ------------ 2.89s",
                                "osism.services.netdata : Add repository gpg key ------------------------- 2.59s",
                                "osism.services.netdata : Manage service netdata ------------------------- 2.47s",
                                "osism.services.netdata : Opt out from anonymous statistics -------------- 2.27s",
                                "osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.87s",
                                "osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.83s",
                                "osism.services.netdata : Remove old architecture-dependent repository --- 1.54s",
                                "osism.services.netdata : Include config tasks --------------------------- 1.40s",
                                "osism.services.netdata : Include host type specific tasks --------------- 1.36s",
                                "osism.services.netdata : Include distribution specific install tasks ---- 1.24s",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fc8d3b04-efac-40a7-bdd7-b8738b782165 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:49:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 06b9be2a-8977-468f-a33d-1b1b7b687823 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role phpmyadmin] ***************************************************",
                                "",
                                "TASK [osism.services.phpmyadmin : Create traefik external network] *************",
                                "Friday 13 March 2026  00:48:03 +0000 (0:00:00.253)       0:00:00.253 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Create required directories] *****************",
                                "Friday 13 March 2026  00:48:05 +0000 (0:00:02.240)       0:00:02.494 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************",
                                "Friday 13 March 2026  00:48:06 +0000 (0:00:01.003)       0:00:03.500 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************",
                                "Friday 13 March 2026  00:48:08 +0000 (0:00:02.043)       0:00:05.543 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******",
                                "Friday 13 March 2026  00:49:22 +0000 (0:01:14.327)       0:01:19.870 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:49:26 +0000 (0:00:03.764)       0:01:23.635 ********** ",
                                "=============================================================================== ",
                                "osism.services.phpmyadmin : Manage phpmyadmin service ------------------ 74.33s",
                                "osism.services.phpmyadmin : Restart phpmyadmin service ------------------ 3.76s",
                                "osism.services.phpmyadmin : Create traefik external network ------------- 2.24s",
                                "osism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 2.04s",
                                "osism.services.phpmyadmin : Create required directories ----------------- 1.01s",
                                "",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Friday 13 March 2026  00:47:34 +0000 (0:00:00.232)       0:00:00.232 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  00:47:35 +0000 (0:00:01.245)       0:00:01.478 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Friday 13 March 2026  00:47:40 +0000 (0:00:04.305)       0:00:05.783 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:01.483)       0:00:07.267 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:47:47 +0000 (0:00:05.518)       0:00:12.785 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Friday 13 March 2026  00:47:50 +0000 (0:00:02.791)       0:00:15.577 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensure /var/log/journal exists on EL10 systems] *****************",
                                "Friday 13 March 2026  00:47:54 +0000 (0:00:04.512)       0:00:20.089 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Friday 13 March 2026  00:47:56 +0000 (0:00:01.742)       0:00:21.832 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Friday 13 March 2026  00:47:58 +0000 (0:00:01.754)       0:00:23.586 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over kolla.target] **************************************",
                                "Friday 13 March 2026  00:47:59 +0000 (0:00:01.285)       0:00:24.872 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  00:48:02 +0000 (0:00:03.573)       0:00:28.446 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Friday 13 March 2026  00:48:10 +0000 (0:00:07.182)       0:00:35.628 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Friday 13 March 2026  00:48:11 +0000 (0:00:01.050)       0:00:36.679 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Friday 13 March 2026  00:48:11 +0000 (0:00:00.744)       0:00:37.423 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Friday 13 March 2026  00:48:12 +0000 (0:00:00.775)       0:00:38.199 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Friday 13 March 2026  00:48:13 +0000 (0:00:00.818)       0:00:39.017 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Friday 13 March 2026  00:48:18 +0000 (0:00:05.011)       0:00:44.029 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Friday 13 March 2026  00:48:21 +0000 (0:00:03.393)       0:00:47.423 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:02.172)       0:00:49.596 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Friday 13 March 2026  00:48:25 +0000 (0:00:01.748)       0:00:51.344 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Friday 13 March 2026  00:48:27 +0000 (0:00:01.896)       0:00:53.241 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Check containers] ********************",
                                "Friday 13 March 2026  00:48:30 +0000 (0:00:02.342)       0:00:55.584 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : common | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:48:33 +0000 (0:00:03.435)       0:00:59.019 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:48:34 +0000 (0:00:01.085)       0:01:00.105 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2025.1', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Creating log volume] ********************************************",
                                "Friday 13 March 2026  00:48:37 +0000 (0:00:02.644)       0:01:02.749 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Link kolla_logs volume to /var/log/kolla] ***********************",
                                "Friday 13 March 2026  00:48:38 +0000 (0:00:01.756)       0:01:04.505 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:01.239)       0:01:05.745 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.089)       0:01:05.834 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.062)       0:01:05.897 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.199)       0:01:06.097 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.059)       0:01:06.156 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.075)       0:01:06.232 ********** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.085)       0:01:06.317 ********** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:00.110)       0:01:06.428 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Friday 13 March 2026  00:49:17 +0000 (0:00:36.775)       0:01:43.204 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:33.262)       0:02:16.467 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Friday 13 March 2026  00:49:52 +0000 (0:00:02.086)       0:02:18.554 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:50:02 +0000 (0:00:09.264)       0:02:27.818 ********** ",
                                "=============================================================================== ",
                                "common : Restart fluentd container ------------------------------------- 36.78s",
                                "common : Restart kolla-toolbox container ------------------------------- 33.26s",
                                "common : Restart cron container ----------------------------------------- 9.26s",
                                "common : Copying over config.json files for services -------------------- 7.18s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 5.52s",
                                "common : Copying over fluentd.conf -------------------------------------- 5.01s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 4.51s",
                                "common : Ensuring config directories exist ------------------------------ 4.31s",
                                "common : Copying over kolla.target -------------------------------------- 3.57s",
                                "service-check-containers : common | Check containers -------------------- 3.44s",
                                "common : Copying over cron logrotate config file ------------------------ 3.39s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 2.79s",
                                "service-check-containers : Include tasks -------------------------------- 2.64s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.34s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.17s",
                                "common : Initializing toolbox container using normal user --------------- 2.09s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 1.90s",
                                "common : Creating log volume -------------------------------------------- 1.76s",
                                "common : Copying over /run subdirectories conf -------------------------- 1.75s",
                                "common : Ensuring config directories have correct owner and permission --- 1.75s",
                                "\u001b[32m2026-03-13 00:50:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.407)       0:00:00.407 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.322)       0:00:00.730 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.466)       0:00:01.196 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.631)       0:00:01.828 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:00.853)       0:00:02.681 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [service-check-containers : memcached | Check containers] *****************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:02.193)       0:00:04.875 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : memcached | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:50:15 +0000 (0:00:01.419)       0:00:06.295 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:50:15 +0000 (0:00:00.455)       0:00:06.750 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Friday 13 March 2026  00:50:17 +0000 (0:00:01.801)       0:00:08.551 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:50:24 +0000 (0:00:07.280)       0:00:15.832 ********** ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container --------------------------------- 7.28s",
                                "memcached : Copying over config.json files for services ----------------- 2.19s",
                                "service-check-containers : Include tasks -------------------------------- 1.80s",
                                "service-check-containers : memcached | Check containers ----------------- 1.42s",
                                "memcached : Ensuring config directories exist --------------------------- 0.85s",
                                "memcached : include_tasks ----------------------------------------------- 0.63s",
                                "Group hosts based on enabled services ----------------------------------- 0.47s",
                                "service-check-containers : memcached | Notify handlers to restart containers --- 0.46s",
                                "Group hosts based on Kolla action --------------------------------------- 0.32s",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 773816ff-4482-48c7-a830-e1cb36c8b52e is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:50:08 +0000 (0:00:00.290)       0:00:00.290 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.381)       0:00:00.671 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.564)       0:00:01.236 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.717)       0:00:01.953 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:01.412)       0:00:03.365 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Friday 13 March 2026  00:50:14 +0000 (0:00:02.709)       0:00:06.075 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : redis | Check containers] *********************",
                                "Friday 13 March 2026  00:50:17 +0000 (0:00:02.917)       0:00:08.993 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : redis | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:50:19 +0000 (0:00:01.852)       0:00:10.846 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:50:19 +0000 (0:00:00.469)       0:00:11.315 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2025.1', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2025.1', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Friday 13 March 2026  00:50:21 +0000 (0:00:01.112)       0:00:12.427 ********** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Friday 13 March 2026  00:50:21 +0000 (0:00:00.112)       0:00:12.540 ********** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Friday 13 March 2026  00:50:21 +0000 (0:00:00.175)       0:00:12.716 ********** ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Friday 13 March 2026  00:50:21 +0000 (0:00:00.124)       0:00:12.840 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Friday 13 March 2026  00:50:24 +0000 (0:00:02.758)       0:00:15.599 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:50:28 +0000 (0:00:04.700)       0:00:20.299 ********** ",
                                "=============================================================================== ",
                                "redis : Restart redis-sentinel container -------------------------------- 4.70s",
                                "redis : Copying over redis config files --------------------------------- 2.92s",
                                "redis : Restart redis container ----------------------------------------- 2.76s",
                                "redis : Copying over default config.json files -------------------------- 2.71s",
                                "service-check-containers : redis | Check containers --------------------- 1.85s",
                                "redis : Ensuring config directories exist ------------------------------- 1.41s",
                                "service-check-containers : Include tasks -------------------------------- 1.11s",
                                "redis : include_tasks --------------------------------------------------- 0.72s",
                                "Group hosts based on enabled services ----------------------------------- 0.56s",
                                "service-check-containers : redis | Notify handlers to restart containers --- 0.47s",
                                "redis : Flush handlers -------------------------------------------------- 0.41s",
                                "Group hosts based on Kolla action --------------------------------------- 0.38s",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb96d8e5-d5e4-4b68-8729-e4165c78fe2b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:50:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 45cdcbfb-461b-4191-89da-b98f2402463f is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:50:08 +0000 (0:00:00.226)       0:00:00.226 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.995)       0:00:01.221 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.670)       0:00:01.892 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:01.360)       0:00:03.253 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:01.460)       0:00:04.714 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Friday 13 March 2026  00:50:15 +0000 (0:00:01.901)       0:00:06.615 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Friday 13 March 2026  00:50:16 +0000 (0:00:01.774)       0:00:08.390 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Friday 13 March 2026  00:50:18 +0000 (0:00:01.588)       0:00:09.978 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Friday 13 March 2026  00:50:20 +0000 (0:00:01.693)       0:00:11.672 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Friday 13 March 2026  00:50:22 +0000 (0:00:02.569)       0:00:14.242 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-check-containers : openvswitch | Check containers] ***************",
                                "Friday 13 March 2026  00:50:23 +0000 (0:00:00.835)       0:00:15.078 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : openvswitch | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:50:26 +0000 (0:00:03.176)       0:00:18.255 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:50:27 +0000 (0:00:01.014)       0:00:19.269 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:02.418)       0:00:21.687 ********** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:00.288)       0:00:21.976 ********** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:00.195)       0:00:22.171 ********** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:00.156)       0:00:22.328 ********** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.486)       0:00:22.814 ********** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.198)       0:00:23.012 ********** ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.169)       0:00:23.182 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:05.232)       0:00:28.414 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Friday 13 March 2026  00:50:38 +0000 (0:00:01.571)       0:00:29.986 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Friday 13 March 2026  00:50:47 +0000 (0:00:08.721)       0:00:38.707 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Friday 13 March 2026  00:50:55 +0000 (0:00:07.797)       0:00:46.504 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Friday 13 March 2026  00:50:57 +0000 (0:00:02.081)       0:00:48.586 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Friday 13 March 2026  00:51:00 +0000 (0:00:03.552)       0:00:52.138 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:51:09 +0000 (0:00:08.398)       0:01:00.537 ********** ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 17.12s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 7.80s",
                                "openvswitch : Restart openvswitch-db-server container ------------------- 5.23s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 3.55s",
                                "service-check-containers : openvswitch | Check containers --------------- 3.18s",
                                "openvswitch : Copying over config.json files for services --------------- 2.57s",
                                "service-check-containers : Include tasks -------------------------------- 2.42s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 2.08s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.90s",
                                "module-load : Drop module persistence ----------------------------------- 1.77s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.69s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 1.59s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.57s",
                                "openvswitch : Flush Handlers -------------------------------------------- 1.49s",
                                "module-load : Load modules ---------------------------------------------- 1.46s",
                                "openvswitch : include_tasks --------------------------------------------- 1.36s",
                                "service-check-containers : openvswitch | Notify handlers to restart containers --- 1.01s",
                                "Group hosts based on Kolla action --------------------------------------- 1.00s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.84s",
                                "Group hosts based on enabled services ----------------------------------- 0.67s",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 784863f2-29ad-4e4f-b185-4c3dc766a5da is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Friday 13 March 2026  00:47:34 +0000 (0:00:00.216)       0:00:00.216 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Friday 13 March 2026  00:47:35 +0000 (0:00:00.696)       0:00:00.912 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Friday 13 March 2026  00:47:36 +0000 (0:00:00.567)       0:00:01.480 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Friday 13 March 2026  00:47:36 +0000 (0:00:00.636)       0:00:02.116 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Friday 13 March 2026  00:47:39 +0000 (0:00:02.605)       0:00:04.722 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:02.356)       0:00:07.078 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Friday 13 March 2026  00:47:42 +0000 (0:00:01.101)       0:00:08.179 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Friday 13 March 2026  00:47:43 +0000 (0:00:00.911)       0:00:09.091 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Friday 13 March 2026  00:47:44 +0000 (0:00:00.773)       0:00:09.865 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Friday 13 March 2026  00:47:45 +0000 (0:00:00.928)       0:00:10.794 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Friday 13 March 2026  00:47:47 +0000 (0:00:01.644)       0:00:12.438 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Friday 13 March 2026  00:47:48 +0000 (0:00:01.069)       0:00:13.508 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Friday 13 March 2026  00:47:53 +0000 (0:00:05.213)       0:00:18.721 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Friday 13 March 2026  00:47:55 +0000 (0:00:02.053)       0:00:20.775 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Friday 13 March 2026  00:47:57 +0000 (0:00:02.361)       0:00:23.137 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Friday 13 March 2026  00:47:59 +0000 (0:00:01.139)       0:00:24.277 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Friday 13 March 2026  00:48:00 +0000 (0:00:01.641)       0:00:25.919 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Friday 13 March 2026  00:48:02 +0000 (0:00:01.477)       0:00:27.396 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Friday 13 March 2026  00:48:04 +0000 (0:00:02.600)       0:00:29.996 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Friday 13 March 2026  00:48:06 +0000 (0:00:01.977)       0:00:31.974 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Friday 13 March 2026  00:48:08 +0000 (0:00:01.464)       0:00:33.438 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Friday 13 March 2026  00:48:09 +0000 (0:00:01.119)       0:00:34.558 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Friday 13 March 2026  00:48:10 +0000 (0:00:00.792)       0:00:35.351 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Friday 13 March 2026  00:48:10 +0000 (0:00:00.312)       0:00:35.664 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Friday 13 March 2026  00:48:11 +0000 (0:00:01.312)       0:00:36.976 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Friday 13 March 2026  00:48:13 +0000 (0:00:01.305)       0:00:38.282 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Friday 13 March 2026  00:48:13 +0000 (0:00:00.455)       0:00:38.738 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Friday 13 March 2026  00:48:16 +0000 (0:00:02.938)       0:00:41.676 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Friday 13 March 2026  00:48:17 +0000 (0:00:00.817)       0:00:42.494 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Friday 13 March 2026  00:48:18 +0000 (0:00:00.886)       0:00:43.381 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Friday 13 March 2026  00:48:19 +0000 (0:00:01.286)       0:00:44.668 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Friday 13 March 2026  00:48:20 +0000 (0:00:00.622)       0:00:45.291 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Friday 13 March 2026  00:48:20 +0000 (0:00:00.375)       0:00:45.666 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Friday 13 March 2026  00:48:21 +0000 (0:00:01.524)       0:00:47.191 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:02.395)       0:00:49.587 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:00.530)       0:00:50.118 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Friday 13 March 2026  00:49:08 +0000 (0:00:43.418)       0:01:33.537 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Friday 13 March 2026  00:49:08 +0000 (0:00:00.341)       0:01:33.879 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Friday 13 March 2026  00:49:09 +0000 (0:00:00.912)       0:01:34.791 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Friday 13 March 2026  00:49:10 +0000 (0:00:01.367)       0:01:36.158 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Friday 13 March 2026  00:49:37 +0000 (0:00:26.118)       0:02:02.277 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Friday 13 March 2026  00:49:37 +0000 (0:00:00.706)       0:02:02.983 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Friday 13 March 2026  00:49:38 +0000 (0:00:00.651)       0:02:03.634 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Friday 13 March 2026  00:49:38 +0000 (0:00:00.551)       0:02:04.186 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Friday 13 March 2026  00:49:39 +0000 (0:00:00.893)       0:02:05.079 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Friday 13 March 2026  00:49:40 +0000 (0:00:00.286)       0:02:05.366 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Friday 13 March 2026  00:49:40 +0000 (0:00:00.638)       0:02:06.005 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Friday 13 March 2026  00:49:41 +0000 (0:00:00.686)       0:02:06.691 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Friday 13 March 2026  00:49:42 +0000 (0:00:01.194)       0:02:07.886 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Friday 13 March 2026  00:49:43 +0000 (0:00:01.049)       0:02:08.935 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Friday 13 March 2026  00:49:43 +0000 (0:00:00.276)       0:02:09.211 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Friday 13 March 2026  00:49:44 +0000 (0:00:00.269)       0:02:09.480 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Friday 13 March 2026  00:49:45 +0000 (0:00:00.858)       0:02:10.338 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Friday 13 March 2026  00:49:45 +0000 (0:00:00.616)       0:02:10.955 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Friday 13 March 2026  00:49:48 +0000 (0:00:02.758)       0:02:13.714 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Friday 13 March 2026  00:49:48 +0000 (0:00:00.523)       0:02:14.237 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Friday 13 March 2026  00:49:49 +0000 (0:00:00.590)       0:02:14.828 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Friday 13 March 2026  00:49:49 +0000 (0:00:00.283)       0:02:15.112 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:00.542)       0:02:15.655 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:00.258)       0:02:15.913 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:00.259)       0:02:16.173 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Friday 13 March 2026  00:49:51 +0000 (0:00:00.324)       0:02:16.497 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Friday 13 March 2026  00:49:52 +0000 (0:00:01.015)       0:02:17.512 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Friday 13 March 2026  00:49:53 +0000 (0:00:01.016)       0:02:18.528 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Friday 13 March 2026  00:49:54 +0000 (0:00:01.320)       0:02:19.848 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Friday 13 March 2026  00:50:04 +0000 (0:00:09.792)       0:02:29.640 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Friday 13 March 2026  00:50:05 +0000 (0:00:00.696)       0:02:30.337 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Friday 13 March 2026  00:50:05 +0000 (0:00:00.500)       0:02:30.838 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Friday 13 March 2026  00:50:06 +0000 (0:00:00.585)       0:02:31.423 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Friday 13 March 2026  00:50:07 +0000 (0:00:00.840)       0:02:32.264 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Friday 13 March 2026  00:50:07 +0000 (0:00:00.984)       0:02:33.248 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:01.256)       0:02:34.505 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.657)       0:02:35.162 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.472)       0:02:35.635 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.346)       0:02:35.982 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.127)       0:02:36.109 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:00.207)       0:02:36.316 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:00.674)       0:02:36.990 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:01.407)       0:02:38.398 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:00.802)       0:02:39.201 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Friday 13 March 2026  00:50:14 +0000 (0:00:00.514)       0:02:39.716 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Friday 13 March 2026  00:50:20 +0000 (0:00:06.379)       0:02:46.095 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Friday 13 March 2026  00:50:34 +0000 (0:00:13.686)       0:02:59.781 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.484)       0:03:00.266 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.283)       0:03:00.549 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.325)       0:03:00.874 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Friday 13 March 2026  00:50:36 +0000 (0:00:00.574)       0:03:01.449 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Friday 13 March 2026  00:50:36 +0000 (0:00:00.781)       0:03:02.231 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:00.798)       0:03:03.029 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:00.103)       0:03:03.132 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Friday 13 March 2026  00:50:38 +0000 (0:00:00.967)       0:03:04.100 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Friday 13 March 2026  00:50:38 +0000 (0:00:00.115)       0:03:04.215 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Friday 13 March 2026  00:50:39 +0000 (0:00:00.112)       0:03:04.327 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Friday 13 March 2026  00:50:39 +0000 (0:00:00.114)       0:03:04.442 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Friday 13 March 2026  00:50:39 +0000 (0:00:00.120)       0:03:04.563 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Friday 13 March 2026  00:50:43 +0000 (0:00:04.216)       0:03:08.779 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Friday 13 March 2026  00:51:30 +0000 (0:00:46.684)       0:03:55.464 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Friday 13 March 2026  00:51:31 +0000 (0:00:01.129)       0:03:56.594 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Friday 13 March 2026  00:51:32 +0000 (0:00:01.597)       0:03:58.191 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Friday 13 March 2026  00:51:33 +0000 (0:00:00.904)       0:03:59.096 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Friday 13 March 2026  00:51:33 +0000 (0:00:00.112)       0:03:59.208 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Friday 13 March 2026  00:51:35 +0000 (0:00:01.844)       0:04:01.052 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Friday 13 March 2026  00:51:36 +0000 (0:00:00.278)       0:04:01.331 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Friday 13 March 2026  00:51:37 +0000 (0:00:01.036)       0:04:02.368 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Friday 13 March 2026  00:51:37 +0000 (0:00:00.121)       0:04:02.489 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Friday 13 March 2026  00:51:37 +0000 (0:00:00.207)       0:04:02.696 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Friday 13 March 2026  00:51:42 +0000 (0:00:05.237)       0:04:07.933 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Friday 13 March 2026  00:51:43 +0000 (0:00:00.774)       0:04:08.708 ********** ",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Friday 13 March 2026  00:51:53 +0000 (0:00:09.607)       0:04:18.315 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Friday 13 March 2026  00:51:53 +0000 (0:00:00.703)       0:04:19.019 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:51:54 +0000 (0:00:00.459)       0:04:19.478 ********** ",
                                "=============================================================================== ",
                                "k3s_server_post : Wait for Cilium resources ---------------------------- 46.68s",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 43.42s",
                                "k3s_server : Enable and check K3s service ------------------------------ 26.12s",
                                "kubectl : Install required packages ------------------------------------ 13.69s",
                                "k3s_agent : Manage k3s service ------------------------------------------ 9.79s",
                                "Manage labels ----------------------------------------------------------- 9.61s",
                                "kubectl : Add repository Debian ----------------------------------------- 6.38s",
                                "k9s : Install k9s packages ---------------------------------------------- 5.24s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 5.21s",
                                "k3s_server_post : Install Cilium ---------------------------------------- 4.22s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.94s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 2.76s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.61s",
                                "k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured --- 2.60s",
                                "k3s_server : Detect Kubernetes version for label compatibility ---------- 2.40s",
                                "k3s_download : Download k3s binary armhf -------------------------------- 2.36s",
                                "k3s_prereq : Enable IPv6 forwarding ------------------------------------- 2.36s",
                                "k3s_download : Download k3s binary arm64 -------------------------------- 2.05s",
                                "k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers --- 1.98s",
                                "k3s_server_post : Test for BGP config resources ------------------------- 1.84s",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5dd18988-6afa-48f6-b9a5-301792612dd3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:51:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5dd18988-6afa-48f6-b9a5-301792612dd3 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4fdf214c-ff1c-400a-8e25-1777494b1cb9 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:52:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7d560e95-76a5-42f1-816e-2c816ce05d1f is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy kubeconfig to the configuration repository] *************************",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Friday 13 March 2026  00:51:58 +0000 (0:00:00.141)       0:00:00.141 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Friday 13 March 2026  00:51:59 +0000 (0:00:00.634)       0:00:00.775 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig file] ****************************",
                                "Friday 13 March 2026  00:52:00 +0000 (0:00:01.017)       0:00:01.793 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:52:00 +0000 (0:00:00.437)       0:00:02.230 ********** ",
                                "=============================================================================== ",
                                "Write kubeconfig file --------------------------------------------------- 1.02s",
                                "Get kubeconfig file ----------------------------------------------------- 0.63s",
                                "Change server address in the kubeconfig file ---------------------------- 0.44s",
                                "",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Friday 13 March 2026  00:51:58 +0000 (0:00:00.119)       0:00:00.119 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Friday 13 March 2026  00:51:58 +0000 (0:00:00.430)       0:00:00.550 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Friday 13 March 2026  00:51:59 +0000 (0:00:00.506)       0:00:01.056 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Friday 13 March 2026  00:52:00 +0000 (0:00:00.655)       0:00:01.712 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Friday 13 March 2026  00:52:01 +0000 (0:00:01.218)       0:00:02.931 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Friday 13 March 2026  00:52:01 +0000 (0:00:00.509)       0:00:03.440 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Friday 13 March 2026  00:52:03 +0000 (0:00:01.937)       0:00:05.378 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Friday 13 March 2026  00:52:04 +0000 (0:00:01.081)       0:00:06.459 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Friday 13 March 2026  00:52:05 +0000 (0:00:00.414)       0:00:06.874 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:52:05 +0000 (0:00:00.328)       0:00:07.202 ********** ",
                                "=============================================================================== ",
                                "Make kubeconfig available for use inside the manager service ------------ 1.94s",
                                "Write kubeconfig file --------------------------------------------------- 1.22s",
                                "Change server address in the kubeconfig inside the manager service ------ 1.08s",
                                "Get kubeconfig file ----------------------------------------------------- 0.66s",
                                "Change server address in the kubeconfig --------------------------------- 0.51s",
                                "Create .kube directory -------------------------------------------------- 0.51s",
                                "Get home directory of operator user ------------------------------------- 0.43s",
                                "Set KUBECONFIG environment variable ------------------------------------- 0.41s",
                                "Enable kubectl command line completion ---------------------------------- 0.33s",
                                "",
                                "",
                                "PLAY [Set kolla_action_rabbitmq] ***********************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.148)       0:00:00.148 ********** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check RabbitMQ service] **************************************************",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.064)       0:00:00.212 ********** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:04.261)       0:00:04.474 ********** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.070)       0:00:04.545 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.243)       0:00:04.788 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:50:36 +0000 (0:00:00.560)       0:00:05.349 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:00.931)       0:00:06.281 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:00.455)       0:00:06.736 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Friday 13 March 2026  00:50:39 +0000 (0:00:01.529)       0:00:08.266 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Friday 13 March 2026  00:50:40 +0000 (0:00:00.957)       0:00:09.224 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Friday 13 March 2026  00:50:41 +0000 (0:00:01.002)       0:00:10.226 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Friday 13 March 2026  00:50:41 +0000 (0:00:00.236)       0:00:10.463 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:50:42 +0000 (0:00:00.646)       0:00:11.109 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Friday 13 March 2026  00:50:42 +0000 (0:00:00.496)       0:00:11.605 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Friday 13 March 2026  00:50:43 +0000 (0:00:00.751)       0:00:12.357 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Friday 13 March 2026  00:50:43 +0000 (0:00:00.277)       0:00:12.635 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Friday 13 March 2026  00:50:44 +0000 (0:00:00.560)       0:00:13.195 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Friday 13 March 2026  00:50:45 +0000 (0:00:00.995)       0:00:14.191 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:01.483)       0:00:15.674 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Friday 13 March 2026  00:50:48 +0000 (0:00:01.506)       0:00:17.180 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Friday 13 March 2026  00:50:50 +0000 (0:00:02.365)       0:00:19.546 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Friday 13 March 2026  00:50:51 +0000 (0:00:01.372)       0:00:20.919 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Friday 13 March 2026  00:50:53 +0000 (0:00:01.788)       0:00:22.708 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Friday 13 March 2026  00:50:54 +0000 (0:00:01.166)       0:00:23.874 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:50:56 +0000 (0:00:01.625)       0:00:25.500 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over extra CA certificates] *******",
                                "Friday 13 March 2026  00:50:57 +0000 (0:00:00.751)       0:00:26.251 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:50:58 +0000 (0:00:01.453)       0:00:27.705 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : rabbitmq | Copying over backend internal TLS key] ****",
                                "Friday 13 March 2026  00:50:59 +0000 (0:00:00.456)       0:00:28.161 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : rabbitmq | Check containers] ******************",
                                "Friday 13 March 2026  00:50:59 +0000 (0:00:00.808)       0:00:28.969 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : rabbitmq | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:51:01 +0000 (0:00:01.165)       0:00:30.135 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:51:01 +0000 (0:00:00.757)       0:00:30.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Creating rabbitmq volume] *************************************",
                                "Friday 13 March 2026  00:51:03 +0000 (0:00:01.760)       0:00:32.653 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Running RabbitMQ bootstrap container] *************************",
                                "Friday 13 March 2026  00:51:04 +0000 (0:00:00.918)       0:00:33.571 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Friday 13 March 2026  00:51:12 +0000 (0:00:07.647)       0:00:41.219 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Friday 13 March 2026  00:51:12 +0000 (0:00:00.741)       0:00:41.961 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Friday 13 March 2026  00:51:13 +0000 (0:00:00.909)       0:00:42.871 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Friday 13 March 2026  00:51:13 +0000 (0:00:00.119)       0:00:42.990 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Friday 13 March 2026  00:51:15 +0000 (0:00:02.008)       0:00:44.998 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Friday 13 March 2026  00:53:08 +0000 (0:01:52.165)       0:02:37.164 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Friday 13 March 2026  00:53:08 +0000 (0:00:00.612)       0:02:37.776 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Friday 13 March 2026  00:53:08 +0000 (0:00:00.115)       0:02:37.892 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Friday 13 March 2026  00:53:10 +0000 (0:00:01.472)       0:02:39.364 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Friday 13 March 2026  00:53:19 +0000 (0:00:09.396)       0:02:48.761 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Friday 13 March 2026  00:53:20 +0000 (0:00:00.567)       0:02:49.328 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Friday 13 March 2026  00:53:20 +0000 (0:00:00.087)       0:02:49.416 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Friday 13 March 2026  00:53:21 +0000 (0:00:01.422)       0:02:50.839 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Friday 13 March 2026  00:53:30 +0000 (0:00:08.948)       0:02:59.787 ********** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Friday 13 March 2026  00:53:31 +0000 (0:00:00.658)       0:03:00.445 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=10  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:53:34 +0000 (0:00:03.283)       0:03:03.729 ********** ",
                                "=============================================================================== ",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------ 130.51s",
                                "rabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.65s",
                                "rabbitmq : Restart rabbitmq container ----------------------------------- 4.90s",
                                "Check RabbitMQ service -------------------------------------------------- 4.26s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 3.28s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 2.37s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 2.09s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 1.79s",
                                "service-check-containers : Include tasks -------------------------------- 1.76s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 1.63s",
                                "rabbitmq : Get container facts ------------------------------------------ 1.53s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.51s",
                                "rabbitmq : Copying over config.json files for services ------------------ 1.48s",
                                "service-cert-copy : rabbitmq | Copying over extra CA certificates ------- 1.45s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 1.37s",
                                "service-check-containers : rabbitmq | Check containers ------------------ 1.17s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 1.17s",
                                "rabbitmq : Get new RabbitMQ version ------------------------------------- 1.00s",
                                "rabbitmq : Ensuring config directories exist ---------------------------- 1.00s",
                                "rabbitmq : Get current RabbitMQ version --------------------------------- 0.96s",
                                "\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:53:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 04e2d29f-14ef-4f4d-b0b8-7866e9f96396 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:51:14 +0000 (0:00:00.151)       0:00:00.151 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:51:15 +0000 (0:00:00.845)       0:00:00.997 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Friday 13 March 2026  00:51:16 +0000 (0:00:00.929)       0:00:01.926 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Friday 13 March 2026  00:51:17 +0000 (0:00:01.100)       0:00:03.026 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Friday 13 March 2026  00:51:18 +0000 (0:00:01.171)       0:00:04.198 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Friday 13 March 2026  00:51:19 +0000 (0:00:01.572)       0:00:05.771 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Friday 13 March 2026  00:51:21 +0000 (0:00:01.243)       0:00:07.014 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_controller | Check containers] ************",
                                "Friday 13 March 2026  00:51:22 +0000 (0:00:01.403)       0:00:08.418 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_controller | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:51:24 +0000 (0:00:01.514)       0:00:09.933 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:51:24 +0000 (0:00:00.896)       0:00:10.829 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Friday 13 March 2026  00:51:25 +0000 (0:00:01.021)       0:00:11.851 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Friday 13 March 2026  00:51:28 +0000 (0:00:02.666)       0:00:14.518 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:16641,tcp:192.168.16.11:16641,tcp:192.168.16.12:16641'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:22.823)       0:00:37.341 ********** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.062)       0:00:37.403 ********** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.081)       0:00:37.485 ********** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.061)       0:00:37.547 ********** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.063)       0:00:37.610 ********** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.074)       0:00:37.684 ********** ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Reload systemd config] ***********************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.080)       0:00:37.764 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Friday 13 March 2026  00:51:54 +0000 (0:00:02.226)       0:00:39.991 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Friday 13 March 2026  00:52:02 +0000 (0:00:08.256)       0:00:48.247 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Friday 13 March 2026  00:52:03 +0000 (0:00:01.088)       0:00:49.336 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Friday 13 March 2026  00:52:04 +0000 (0:00:01.260)       0:00:50.596 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Friday 13 March 2026  00:52:05 +0000 (0:00:00.988)       0:00:51.585 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Friday 13 March 2026  00:52:05 +0000 (0:00:00.286)       0:00:51.871 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Friday 13 March 2026  00:52:06 +0000 (0:00:00.387)       0:00:52.259 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Friday 13 March 2026  00:52:06 +0000 (0:00:00.297)       0:00:52.557 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Friday 13 March 2026  00:52:06 +0000 (0:00:00.272)       0:00:52.829 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Friday 13 March 2026  00:52:07 +0000 (0:00:00.275)       0:00:53.105 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Friday 13 March 2026  00:52:07 +0000 (0:00:00.599)       0:00:53.704 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Friday 13 March 2026  00:52:08 +0000 (0:00:00.375)       0:00:54.080 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Friday 13 March 2026  00:52:08 +0000 (0:00:00.301)       0:00:54.381 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Friday 13 March 2026  00:52:08 +0000 (0:00:00.301)       0:00:54.683 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Friday 13 March 2026  00:52:09 +0000 (0:00:00.305)       0:00:54.988 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Friday 13 March 2026  00:52:09 +0000 (0:00:00.421)       0:00:55.410 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Friday 13 March 2026  00:52:09 +0000 (0:00:00.243)       0:00:55.654 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Friday 13 March 2026  00:52:10 +0000 (0:00:00.258)       0:00:55.912 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Friday 13 March 2026  00:52:10 +0000 (0:00:00.362)       0:00:56.274 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Friday 13 March 2026  00:52:10 +0000 (0:00:00.444)       0:00:56.718 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Friday 13 March 2026  00:52:11 +0000 (0:00:00.298)       0:00:57.017 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************",
                                "Friday 13 March 2026  00:52:11 +0000 (0:00:00.541)       0:00:57.558 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************",
                                "Friday 13 March 2026  00:52:12 +0000 (0:00:00.569)       0:00:58.128 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check NB cluster status] ****************************************",
                                "Friday 13 March 2026  00:52:12 +0000 (0:00:00.387)       0:00:58.515 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check SB cluster status] ****************************************",
                                "Friday 13 March 2026  00:52:12 +0000 (0:00:00.301)       0:00:58.816 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***",
                                "Friday 13 March 2026  00:52:13 +0000 (0:00:00.282)       0:00:59.099 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***",
                                "Friday 13 March 2026  00:52:13 +0000 (0:00:00.426)       0:00:59.525 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************",
                                "Friday 13 March 2026  00:52:13 +0000 (0:00:00.271)       0:00:59.797 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************",
                                "Friday 13 March 2026  00:52:14 +0000 (0:00:00.298)       0:01:00.095 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  00:52:14 +0000 (0:00:00.277)       0:01:00.373 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  00:52:17 +0000 (0:00:02.751)       0:01:03.125 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Ensure configuration for relays exists] *************************",
                                "Friday 13 March 2026  00:52:23 +0000 (0:00:05.768)       0:01:08.893 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****",
                                "Friday 13 March 2026  00:52:23 +0000 (0:00:00.691)       0:01:09.584 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for OVN relay services] **********",
                                "Friday 13 March 2026  00:52:24 +0000 (0:00:00.662)       0:01:10.247 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Generate config files for OVN relay services] *******************",
                                "Friday 13 March 2026  00:52:26 +0000 (0:00:01.828)       0:01:12.075 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers] ********************",
                                "Friday 13 March 2026  00:52:28 +0000 (0:00:01.854)       0:01:13.929 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:04.339)       0:01:18.269 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:00.423)       0:01:18.693 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers with iteration] *****",
                                "Friday 13 March 2026  00:52:35 +0000 (0:00:02.272)       0:01:20.965 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:52:35 +0000 (0:00:00.911)       0:01:21.876 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.580)       0:01:22.457 ********** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.072)       0:01:22.530 ********** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.062)       0:01:22.593 ********** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.052)       0:01:22.645 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Friday 13 March 2026  00:52:44 +0000 (0:00:07.648)       0:01:30.294 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************",
                                "Friday 13 March 2026  00:52:56 +0000 (0:00:11.780)       0:01:42.075 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Friday 13 March 2026  00:53:10 +0000 (0:00:14.549)       0:01:56.624 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Friday 13 March 2026  00:53:24 +0000 (0:00:13.437)       0:02:10.062 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Friday 13 March 2026  00:53:24 +0000 (0:00:00.110)       0:02:10.172 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Friday 13 March 2026  00:53:25 +0000 (0:00:00.753)       0:02:10.926 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Friday 13 March 2026  00:53:25 +0000 (0:00:00.601)       0:02:11.527 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Friday 13 March 2026  00:53:26 +0000 (0:00:01.008)       0:02:12.536 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Friday 13 March 2026  00:53:27 +0000 (0:00:00.552)       0:02:13.088 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Friday 13 March 2026  00:53:27 +0000 (0:00:00.740)       0:02:13.829 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************",
                                "Friday 13 March 2026  00:53:28 +0000 (0:00:00.773)       0:02:14.602 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m",
                                "",
                                "TASK [ovn-db : Unset bootstrap args fact] **************************************",
                                "Friday 13 March 2026  00:53:29 +0000 (0:00:01.092)       0:02:15.694 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  00:53:30 +0000 (0:00:00.316)       0:02:16.011 ********** ",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  00:53:34 +0000 (0:00:03.899)       0:02:19.910 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [ovn-db : Ensure configuration for relays exists] *************************",
                                "Friday 13 March 2026  00:53:39 +0000 (0:00:05.805)       0:02:25.716 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/config-relay.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=1)\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist for OVN relay containers] *****",
                                "Friday 13 March 2026  00:53:40 +0000 (0:00:00.577)       0:02:26.294 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for OVN relay services] **********",
                                "Friday 13 March 2026  00:53:41 +0000 (0:00:00.671)       0:02:26.965 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Generate config files for OVN relay services] *******************",
                                "Friday 13 March 2026  00:53:42 +0000 (0:00:01.553)       0:02:28.519 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers] ********************",
                                "Friday 13 March 2026  00:53:44 +0000 (0:00:01.750)       0:02:30.269 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:53:49 +0000 (0:00:04.798)       0:02:35.068 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:53:49 +0000 (0:00:00.302)       0:02:35.370 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641', 'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-northd:2025.1', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'environment': {'OVN_NB_DB': 'tcp:192.168.16.10:6641,tcp:192.168.16.11:6641,tcp:192.168.16.12:6641'}, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'environment': {'OVN_SB_DB': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2025.1', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/service-check-containers/tasks/iterated.yml for testbed-node-0, testbed-node-2, testbed-node-1 => (item={'key': 'ovn-sb-db-relay', 'value': {'container_name': 'ovn_sb_db_relay', 'group': 'ovn-sb-db-relay', 'enabled': True, 'environment': {'RELAY_ID': '1'}, 'image': 'registry.osism.tech/kolla/ovn-sb-db-relay:2025.1', 'iterate': True, 'iterate_var': '1', 'volumes': ['/etc/kolla/ovn-sb-db-relay/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Check containers with iteration] *****",
                                "Friday 13 March 2026  00:53:51 +0000 (0:00:01.851)       0:02:37.222 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[1])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[1])\u001b[0m",
                                "",
                                "TASK [service-check-containers : ovn_db | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:01.011)       0:02:38.234 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:00.495)       0:02:38.729 ********** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:00.058)       0:02:38.787 ********** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:00.057)       0:02:38.845 ********** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Friday 13 March 2026  00:53:53 +0000 (0:00:00.058)       0:02:38.904 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Friday 13 March 2026  00:54:05 +0000 (0:00:12.287)       0:02:51.191 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db-relay container] *******************",
                                "Friday 13 March 2026  00:54:18 +0000 (0:00:12.841)       0:03:04.032 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=1)\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Friday 13 March 2026  00:54:30 +0000 (0:00:12.650)       0:03:16.683 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Friday 13 March 2026  00:54:30 +0000 (0:00:00.109)       0:03:16.792 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Friday 13 March 2026  00:54:31 +0000 (0:00:00.744)       0:03:17.537 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Friday 13 March 2026  00:54:32 +0000 (0:00:00.594)       0:03:18.132 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Friday 13 March 2026  00:54:33 +0000 (0:00:00.926)       0:03:19.058 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Friday 13 March 2026  00:54:33 +0000 (0:00:00.633)       0:03:19.692 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Friday 13 March 2026  00:54:34 +0000 (0:00:00.723)       0:03:20.416 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db-relay] ***************************************",
                                "Friday 13 March 2026  00:54:35 +0000 (0:00:00.894)       0:03:21.310 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=1)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=65  \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=63  \u001b[0m \u001b[0;33mchanged=30  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=23  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=63  \u001b[0m \u001b[0;33mchanged=30  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=23  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:54:36 +0000 (0:00:01.294)       0:03:22.604 ********** ",
                                "=============================================================================== ",
                                "ovn-db : Restart ovn-sb-db-relay container ----------------------------- 27.20s",
                                "ovn-db : Restart ovn-sb-db container ----------------------------------- 24.62s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 22.82s",
                                "ovn-db : Restart ovn-nb-db container ----------------------------------- 19.94s",
                                "ovn-db : Restart ovn-northd container ---------------------------------- 13.44s",
                                "ovn-controller : Restart ovn-controller container ----------------------- 8.26s",
                                "ovn-db : Copying over config.json files for services -------------------- 5.81s",
                                "ovn-db : Copying over config.json files for services -------------------- 5.77s",
                                "service-check-containers : ovn_db | Check containers -------------------- 4.80s",
                                "service-check-containers : ovn_db | Check containers -------------------- 4.34s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 3.90s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 2.75s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.67s",
                                "service-check-containers : Include tasks -------------------------------- 2.27s",
                                "ovn-controller : Reload systemd config ---------------------------------- 2.23s",
                                "ovn-db : Generate config files for OVN relay services ------------------- 1.85s",
                                "service-check-containers : Include tasks -------------------------------- 1.85s",
                                "ovn-db : Copying over config.json files for OVN relay services ---------- 1.83s",
                                "ovn-db : Generate config files for OVN relay services ------------------- 1.75s",
                                "ovn-controller : Copying over config.json files for services ------------ 1.57s",
                                "\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:54:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:55:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6483c7df-633c-446b-99fe-79768514f279 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.376)       0:00:00.376 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:50:09 +0000 (0:00:00.419)       0:00:00.796 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Friday 13 March 2026  00:50:10 +0000 (0:00:00.782)       0:00:01.579 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Friday 13 March 2026  00:50:11 +0000 (0:00:00.776)       0:00:02.355 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Friday 13 March 2026  00:50:12 +0000 (0:00:00.921)       0:00:03.276 ********** ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:01.030)       0:00:04.307 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Friday 13 March 2026  00:50:14 +0000 (0:00:00.829)       0:00:05.136 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Friday 13 March 2026  00:50:17 +0000 (0:00:02.964)       0:00:08.101 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Friday 13 March 2026  00:50:18 +0000 (0:00:01.138)       0:00:09.239 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Friday 13 March 2026  00:50:19 +0000 (0:00:01.445)       0:00:10.685 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Friday 13 March 2026  00:50:20 +0000 (0:00:01.173)       0:00:11.859 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Friday 13 March 2026  00:50:22 +0000 (0:00:01.952)       0:00:13.811 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Friday 13 March 2026  00:50:23 +0000 (0:00:01.001)       0:00:14.812 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Friday 13 March 2026  00:50:26 +0000 (0:00:02.322)       0:00:17.135 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Friday 13 March 2026  00:50:27 +0000 (0:00:01.306)       0:00:18.442 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:02.811)       0:00:21.253 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.911)       0:00:22.165 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2025.1', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43', '__omit_place_holder__99989db37d47c12faabbd572900c76e1a130dd43'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:04.355)       0:00:26.520 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Friday 13 March 2026  00:50:40 +0000 (0:00:04.744)       0:00:31.265 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Friday 13 March 2026  00:50:42 +0000 (0:00:02.533)       0:00:33.798 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:03.561)       0:00:37.360 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:00.626)       0:00:37.986 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Friday 13 March 2026  00:50:49 +0000 (0:00:02.865)       0:00:40.851 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Friday 13 March 2026  00:50:52 +0000 (0:00:02.256)       0:00:43.108 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Friday 13 March 2026  00:50:52 +0000 (0:00:00.742)       0:00:43.850 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Friday 13 March 2026  00:50:54 +0000 (0:00:01.713)       0:00:45.564 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql-cert.pem] ***************************",
                                "Friday 13 March 2026  00:50:56 +0000 (0:00:02.235)       0:00:47.800 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql-key.pem] ****************************",
                                "Friday 13 March 2026  00:50:57 +0000 (0:00:00.320)       0:00:48.120 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  00:50:57 +0000 (0:00:00.532)       0:00:48.653 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:51:00 +0000 (0:00:02.944)       0:00:51.597 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  00:51:03 +0000 (0:00:02.685)       0:00:54.282 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Friday 13 March 2026  00:51:04 +0000 (0:00:01.113)       0:00:55.396 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Friday 13 March 2026  00:51:05 +0000 (0:00:01.319)       0:00:56.715 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Friday 13 March 2026  00:51:07 +0000 (0:00:01.475)       0:00:58.190 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Check containers] **************",
                                "Friday 13 March 2026  00:51:08 +0000 (0:00:00.848)       0:00:59.039 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:51:10 +0000 (0:00:02.467)       0:01:01.507 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:51:10 +0000 (0:00:00.323)       0:01:01.830 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Friday 13 March 2026  00:51:12 +0000 (0:00:01.272)       0:01:03.103 ********** ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Friday 13 March 2026  00:51:12 +0000 (0:00:00.596)       0:01:03.699 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Friday 13 March 2026  00:51:16 +0000 (0:00:03.953)       0:01:07.653 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2025.1', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2025.1', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2025.1', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2025.1', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Friday 13 March 2026  00:51:17 +0000 (0:00:00.577)       0:01:08.231 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Friday 13 March 2026  00:51:18 +0000 (0:00:00.821)       0:01:09.052 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Friday 13 March 2026  00:51:19 +0000 (0:00:01.377)       0:01:10.430 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Friday 13 March 2026  00:51:21 +0000 (0:00:01.808)       0:01:12.238 ********** ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Friday 13 March 2026  00:51:21 +0000 (0:00:00.608)       0:01:12.847 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Friday 13 March 2026  00:51:25 +0000 (0:00:04.133)       0:01:16.980 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Friday 13 March 2026  00:51:26 +0000 (0:00:00.716)       0:01:17.697 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Friday 13 March 2026  00:51:27 +0000 (0:00:00.889)       0:01:18.586 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Friday 13 March 2026  00:51:28 +0000 (0:00:01.267)       0:01:19.854 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Friday 13 March 2026  00:51:32 +0000 (0:00:03.311)       0:01:23.165 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Friday 13 March 2026  00:51:32 +0000 (0:00:00.273)       0:01:23.439 ********** ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Friday 13 March 2026  00:51:33 +0000 (0:00:00.761)       0:01:24.201 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Friday 13 March 2026  00:51:35 +0000 (0:00:02.677)       0:01:26.878 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Friday 13 March 2026  00:51:37 +0000 (0:00:01.433)       0:01:28.312 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Friday 13 March 2026  00:51:38 +0000 (0:00:01.633)       0:01:29.946 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Friday 13 March 2026  00:51:39 +0000 (0:00:00.430)       0:01:30.376 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Friday 13 March 2026  00:51:40 +0000 (0:00:01.034)       0:01:31.410 ********** ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Friday 13 March 2026  00:51:41 +0000 (0:00:00.778)       0:01:32.189 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Friday 13 March 2026  00:51:44 +0000 (0:00:03.261)       0:01:35.451 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Friday 13 March 2026  00:51:45 +0000 (0:00:00.948)       0:01:36.400 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Friday 13 March 2026  00:51:46 +0000 (0:00:01.329)       0:01:37.729 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Friday 13 March 2026  00:51:48 +0000 (0:00:01.524)       0:01:39.254 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Friday 13 March 2026  00:51:50 +0000 (0:00:01.929)       0:01:41.184 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Friday 13 March 2026  00:51:50 +0000 (0:00:00.232)       0:01:41.416 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Friday 13 March 2026  00:51:50 +0000 (0:00:00.230)       0:01:41.647 ********** ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Friday 13 March 2026  00:51:51 +0000 (0:00:00.718)       0:01:42.365 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Friday 13 March 2026  00:51:55 +0000 (0:00:03.877)       0:01:46.243 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2025.1', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Friday 13 March 2026  00:51:56 +0000 (0:00:00.976)       0:01:47.220 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Friday 13 March 2026  00:51:57 +0000 (0:00:01.297)       0:01:48.518 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:51:58 +0000 (0:00:01.163)       0:01:49.682 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Friday 13 March 2026  00:52:00 +0000 (0:00:01.790)       0:01:51.473 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Friday 13 March 2026  00:52:00 +0000 (0:00:00.273)       0:01:51.746 ********** ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Friday 13 March 2026  00:52:01 +0000 (0:00:00.841)       0:01:52.587 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Friday 13 March 2026  00:52:06 +0000 (0:00:05.205)       0:01:57.793 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2025.1', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Friday 13 March 2026  00:52:09 +0000 (0:00:02.656)       0:02:00.449 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Friday 13 March 2026  00:52:12 +0000 (0:00:02.982)       0:02:03.432 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Friday 13 March 2026  00:52:13 +0000 (0:00:01.317)       0:02:04.749 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Friday 13 March 2026  00:52:15 +0000 (0:00:01.889)       0:02:06.639 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Friday 13 March 2026  00:52:15 +0000 (0:00:00.305)       0:02:06.944 ********** ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Friday 13 March 2026  00:52:16 +0000 (0:00:00.991)       0:02:07.936 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Friday 13 March 2026  00:52:20 +0000 (0:00:03.716)       0:02:11.652 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Friday 13 March 2026  00:52:21 +0000 (0:00:00.380)       0:02:12.033 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Friday 13 March 2026  00:52:21 +0000 (0:00:00.614)       0:02:12.648 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:52:23 +0000 (0:00:01.492)       0:02:14.141 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Friday 13 March 2026  00:52:25 +0000 (0:00:02.372)       0:02:16.513 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Friday 13 March 2026  00:52:25 +0000 (0:00:00.326)       0:02:16.839 ********** ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Friday 13 March 2026  00:52:26 +0000 (0:00:00.917)       0:02:17.756 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Friday 13 March 2026  00:52:31 +0000 (0:00:04.757)       0:02:22.514 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:00.584)       0:02:23.099 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Friday 13 March 2026  00:52:33 +0000 (0:00:00.945)       0:02:24.044 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:52:34 +0000 (0:00:01.763)       0:02:25.808 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:01.767)       0:02:27.576 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.242)       0:02:27.819 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Friday 13 March 2026  00:52:37 +0000 (0:00:00.259)       0:02:28.078 ********** ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Friday 13 March 2026  00:52:38 +0000 (0:00:01.019)       0:02:29.097 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Friday 13 March 2026  00:52:41 +0000 (0:00:03.857)       0:02:32.955 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Friday 13 March 2026  00:52:42 +0000 (0:00:00.684)       0:02:33.640 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Friday 13 March 2026  00:52:43 +0000 (0:00:00.868)       0:02:34.508 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Friday 13 March 2026  00:52:44 +0000 (0:00:01.159)       0:02:35.667 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Friday 13 March 2026  00:52:46 +0000 (0:00:01.845)       0:02:37.513 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Friday 13 March 2026  00:52:46 +0000 (0:00:00.319)       0:02:37.833 ********** ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Friday 13 March 2026  00:52:48 +0000 (0:00:01.420)       0:02:39.254 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Friday 13 March 2026  00:52:52 +0000 (0:00:03.861)       0:02:43.115 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Friday 13 March 2026  00:52:52 +0000 (0:00:00.560)       0:02:43.676 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Friday 13 March 2026  00:52:53 +0000 (0:00:00.801)       0:02:44.477 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Friday 13 March 2026  00:52:54 +0000 (0:00:01.274)       0:02:45.751 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Friday 13 March 2026  00:52:56 +0000 (0:00:01.932)       0:02:47.684 ********** ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Friday 13 March 2026  00:52:57 +0000 (0:00:01.014)       0:02:48.698 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Friday 13 March 2026  00:53:02 +0000 (0:00:05.110)       0:02:53.808 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Friday 13 March 2026  00:53:03 +0000 (0:00:01.047)       0:02:54.856 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Friday 13 March 2026  00:53:04 +0000 (0:00:00.763)       0:02:55.620 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Friday 13 March 2026  00:53:05 +0000 (0:00:01.219)       0:02:56.839 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Friday 13 March 2026  00:53:07 +0000 (0:00:01.788)       0:02:58.627 ********** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Friday 13 March 2026  00:53:08 +0000 (0:00:01.102)       0:02:59.730 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Friday 13 March 2026  00:53:11 +0000 (0:00:03.167)       0:03:02.897 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Friday 13 March 2026  00:53:15 +0000 (0:00:03.348)       0:03:06.246 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2025.1', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Friday 13 March 2026  00:53:17 +0000 (0:00:02.755)       0:03:09.002 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Friday 13 March 2026  00:53:19 +0000 (0:00:01.864)       0:03:10.867 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:53:21 +0000 (0:00:01.621)       0:03:12.488 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Friday 13 March 2026  00:53:22 +0000 (0:00:01.258)       0:03:13.746 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Friday 13 March 2026  00:53:22 +0000 (0:00:00.261)       0:03:14.008 ********** ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Friday 13 March 2026  00:53:23 +0000 (0:00:00.933)       0:03:14.941 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Friday 13 March 2026  00:53:25 +0000 (0:00:01.332)       0:03:16.273 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2025.1', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Friday 13 March 2026  00:53:25 +0000 (0:00:00.378)       0:03:16.652 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Friday 13 March 2026  00:53:26 +0000 (0:00:00.849)       0:03:17.502 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:53:26 +0000 (0:00:00.418)       0:03:17.920 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Friday 13 March 2026  00:53:28 +0000 (0:00:01.216)       0:03:19.137 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Friday 13 March 2026  00:53:28 +0000 (0:00:00.300)       0:03:19.437 ********** ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Friday 13 March 2026  00:53:29 +0000 (0:00:01.397)       0:03:20.835 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Friday 13 March 2026  00:53:35 +0000 (0:00:05.803)       0:03:26.638 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2025.1', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}, 'pid_mode': '', 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-dhcp-agent:2025.1', 'KOLLA_NAME': 'neutron_dhcp_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'cgroupns_mode': 'private', 'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_IMAGE': 'registry.osism.tech/kolla/neutron-l3-agent:2025.1', 'KOLLA_LEGACY_IPTABLES': 'false', 'KOLLA_NAME': 'neutron_l3_agent', 'KOLLA_NEUTRON_WRAPPERS': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}, 'pid_mode': ''}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2025.1', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': 'NONE', 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2025.1', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2025.1', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2025.1', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2025.1', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Friday 13 March 2026  00:53:37 +0000 (0:00:02.158)       0:03:28.796 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Friday 13 March 2026  00:53:39 +0000 (0:00:01.506)       0:03:30.303 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:53:40 +0000 (0:00:01.208)       0:03:31.512 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Friday 13 March 2026  00:53:42 +0000 (0:00:02.130)       0:03:33.643 ********** ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Friday 13 March 2026  00:53:44 +0000 (0:00:01.379)       0:03:35.022 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Friday 13 March 2026  00:53:47 +0000 (0:00:03.946)       0:03:38.969 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Friday 13 March 2026  00:53:48 +0000 (0:00:00.430)       0:03:39.399 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Friday 13 March 2026  00:53:49 +0000 (0:00:00.885)       0:03:40.285 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:53:50 +0000 (0:00:01.130)       0:03:41.415 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:01.857)       0:03:43.273 ********** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Friday 13 March 2026  00:53:53 +0000 (0:00:01.168)       0:03:44.442 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:00 +0000 (0:00:06.585)       0:03:51.028 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2025.1', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Friday 13 March 2026  00:54:00 +0000 (0:00:00.680)       0:03:51.708 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Friday 13 March 2026  00:54:01 +0000 (0:00:00.813)       0:03:52.521 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Friday 13 March 2026  00:54:02 +0000 (0:00:01.306)       0:03:53.828 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Friday 13 March 2026  00:54:04 +0000 (0:00:01.893)       0:03:55.722 ********** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Friday 13 March 2026  00:54:06 +0000 (0:00:01.395)       0:03:57.117 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Friday 13 March 2026  00:54:07 +0000 (0:00:01.488)       0:03:58.606 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:12 +0000 (0:00:04.675)       0:04:03.282 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Friday 13 March 2026  00:54:13 +0000 (0:00:01.107)       0:04:04.390 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Friday 13 March 2026  00:54:14 +0000 (0:00:01.559)       0:04:05.949 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:54:17 +0000 (0:00:02.171)       0:04:08.120 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Friday 13 March 2026  00:54:20 +0000 (0:00:03.231)       0:04:11.352 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-2, testbed-node-0, testbed-node-1 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Friday 13 March 2026  00:54:21 +0000 (0:00:00.860)       0:04:12.213 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:22 +0000 (0:00:01.679)       0:04:13.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Friday 13 March 2026  00:54:24 +0000 (0:00:01.938)       0:04:15.831 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Friday 13 March 2026  00:54:25 +0000 (0:00:01.038)       0:04:16.870 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:54:27 +0000 (0:00:01.980)       0:04:18.850 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Friday 13 March 2026  00:54:30 +0000 (0:00:02.660)       0:04:21.511 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Friday 13 March 2026  00:54:31 +0000 (0:00:00.756)       0:04:22.267 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:32 +0000 (0:00:01.390)       0:04:23.657 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Friday 13 March 2026  00:54:33 +0000 (0:00:01.218)       0:04:24.876 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Friday 13 March 2026  00:54:35 +0000 (0:00:01.458)       0:04:26.334 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Friday 13 March 2026  00:54:37 +0000 (0:00:02.401)       0:04:28.736 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Friday 13 March 2026  00:54:40 +0000 (0:00:03.110)       0:04:31.847 ********** ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Friday 13 March 2026  00:54:42 +0000 (0:00:01.276)       0:04:33.123 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:45 +0000 (0:00:03.645)       0:04:36.769 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Friday 13 March 2026  00:54:46 +0000 (0:00:00.961)       0:04:37.730 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Friday 13 March 2026  00:54:47 +0000 (0:00:00.808)       0:04:38.539 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:54:48 +0000 (0:00:01.268)       0:04:39.808 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Friday 13 March 2026  00:54:50 +0000 (0:00:01.919)       0:04:41.727 ********** ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Friday 13 March 2026  00:54:52 +0000 (0:00:01.403)       0:04:43.130 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Friday 13 March 2026  00:54:56 +0000 (0:00:04.654)       0:04:47.785 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Friday 13 March 2026  00:54:57 +0000 (0:00:00.828)       0:04:48.614 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Friday 13 March 2026  00:54:58 +0000 (0:00:01.103)       0:04:49.717 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Friday 13 March 2026  00:54:59 +0000 (0:00:00.383)       0:04:50.101 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Friday 13 March 2026  00:55:00 +0000 (0:00:01.277)       0:04:51.378 ********** ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Friday 13 March 2026  00:55:02 +0000 (0:00:01.689)       0:04:53.068 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Friday 13 March 2026  00:55:06 +0000 (0:00:04.417)       0:04:57.485 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['option httpchk', 'timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Friday 13 March 2026  00:55:07 +0000 (0:00:00.825)       0:04:58.310 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Friday 13 March 2026  00:55:08 +0000 (0:00:01.258)       0:04:59.569 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Friday 13 March 2026  00:55:09 +0000 (0:00:00.467)       0:05:00.036 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Friday 13 March 2026  00:55:10 +0000 (0:00:01.310)       0:05:01.347 ********** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Friday 13 March 2026  00:55:11 +0000 (0:00:01.410)       0:05:02.757 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Friday 13 March 2026  00:55:14 +0000 (0:00:02.433)       0:05:05.190 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2025.1', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Friday 13 March 2026  00:55:14 +0000 (0:00:00.464)       0:05:05.655 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Friday 13 March 2026  00:55:15 +0000 (0:00:00.620)       0:05:06.276 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Friday 13 March 2026  00:55:16 +0000 (0:00:00.847)       0:05:07.124 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Friday 13 March 2026  00:55:17 +0000 (0:00:01.297)       0:05:08.422 ********** ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Friday 13 March 2026  00:55:18 +0000 (0:00:01.499)       0:05:09.921 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Friday 13 March 2026  00:55:24 +0000 (0:00:05.922)       0:05:15.844 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2025.1', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2025.1', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Friday 13 March 2026  00:55:25 +0000 (0:00:01.134)       0:05:16.978 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /docs']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Friday 13 March 2026  00:55:26 +0000 (0:00:00.942)       0:05:17.920 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Friday 13 March 2026  00:55:28 +0000 (0:00:01.132)       0:05:19.053 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Friday 13 March 2026  00:55:30 +0000 (0:00:01.977)       0:05:21.030 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Friday 13 March 2026  00:55:30 +0000 (0:00:00.603)       0:05:21.634 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Friday 13 March 2026  00:55:30 +0000 (0:00:00.332)       0:05:21.966 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Friday 13 March 2026  00:55:31 +0000 (0:00:00.298)       0:05:22.265 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Friday 13 March 2026  00:55:31 +0000 (0:00:00.312)       0:05:22.577 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : loadbalancer] *********************************************",
                                "Friday 13 March 2026  00:55:31 +0000 (0:00:00.299)       0:05:22.877 ********** ",
                                "\u001b[0;36mincluded: loadbalancer for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Check containers] **************",
                                "Friday 13 March 2026  00:55:33 +0000 (0:00:01.781)       0:05:24.658 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : loadbalancer | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:55:36 +0000 (0:00:02.571)       0:05:27.230 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:55:36 +0000 (0:00:00.663)       0:05:27.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2025.1', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Friday 13 March 2026  00:55:38 +0000 (0:00:01.529)       0:05:29.423 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Friday 13 March 2026  00:55:39 +0000 (0:00:00.675)       0:05:30.099 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Friday 13 March 2026  00:55:39 +0000 (0:00:00.310)       0:05:30.410 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Friday 13 March 2026  00:55:40 +0000 (0:00:00.744)       0:05:31.154 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Friday 13 March 2026  00:55:41 +0000 (0:00:01.073)       0:05:32.227 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Friday 13 March 2026  00:55:42 +0000 (0:00:00.850)       0:05:33.078 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Friday 13 March 2026  00:55:50 +0000 (0:00:08.898)       0:05:41.976 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Friday 13 March 2026  00:55:51 +0000 (0:00:00.870)       0:05:42.847 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Friday 13 March 2026  00:55:59 +0000 (0:00:08.109)       0:05:50.957 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Friday 13 March 2026  00:56:03 +0000 (0:00:03.106)       0:05:54.063 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Friday 13 March 2026  00:56:07 +0000 (0:00:04.401)       0:05:58.464 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Friday 13 March 2026  00:56:07 +0000 (0:00:00.353)       0:05:58.818 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Friday 13 March 2026  00:56:08 +0000 (0:00:00.352)       0:05:59.171 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Friday 13 March 2026  00:56:08 +0000 (0:00:00.681)       0:05:59.852 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Friday 13 March 2026  00:56:09 +0000 (0:00:00.357)       0:06:00.210 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Friday 13 March 2026  00:56:09 +0000 (0:00:00.365)       0:06:00.575 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Friday 13 March 2026  00:56:09 +0000 (0:00:00.344)       0:06:00.920 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Friday 13 March 2026  00:56:15 +0000 (0:00:05.096)       0:06:06.016 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=126 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=126 \u001b[0m \u001b[0;33mchanged=79  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=94  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:56:15 +0000 (0:00:00.918)       0:06:06.935 ********** ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup haproxy container --------------------------- 8.90s",
                                "loadbalancer : Start backup proxysql container -------------------------- 8.11s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 6.59s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 5.92s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 5.80s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 5.21s",
                                "haproxy-config : Copying over manila haproxy config --------------------- 5.11s",
                                "loadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.10s",
                                "haproxy-config : Copying over horizon haproxy config -------------------- 4.76s",
                                "loadbalancer : Copying over config.json files for services -------------- 4.74s",
                                "haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.68s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 4.65s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 4.42s",
                                "loadbalancer : Start backup keepalived container ------------------------ 4.40s",
                                "loadbalancer : Copying checks for services which are enabled ------------ 4.36s",
                                "haproxy-config : Copying over barbican haproxy config ------------------- 4.13s",
                                "haproxy-config : Copying over aodh haproxy config ----------------------- 3.95s",
                                "haproxy-config : Copying over placement haproxy config ------------------ 3.95s",
                                "haproxy-config : Copying over designate haproxy config ------------------ 3.88s",
                                "haproxy-config : Copying over magnum haproxy config --------------------- 3.86s",
                                "\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d988ebd8-d7ba-479a-9f22-a7a42b5fd790 is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Prepare deployment of Ceph services] *************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Friday 13 March 2026  00:47:35 +0000 (0:00:00.660)       0:00:00.660 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Friday 13 March 2026  00:47:36 +0000 (0:00:01.083)       0:00:01.743 ********** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Friday 13 March 2026  00:47:37 +0000 (0:00:01.677)       0:00:03.421 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Friday 13 March 2026  00:47:38 +0000 (0:00:01.040)       0:00:04.461 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Friday 13 March 2026  00:47:39 +0000 (0:00:01.044)       0:00:05.506 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Friday 13 March 2026  00:47:40 +0000 (0:00:00.851)       0:00:06.358 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Friday 13 March 2026  00:47:41 +0000 (0:00:00.633)       0:00:06.992 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Friday 13 March 2026  00:47:42 +0000 (0:00:01.136)       0:00:08.128 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Friday 13 March 2026  00:47:43 +0000 (0:00:00.777)       0:00:08.905 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Friday 13 March 2026  00:47:44 +0000 (0:00:00.906)       0:00:09.812 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Friday 13 March 2026  00:47:44 +0000 (0:00:00.471)       0:00:10.283 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Friday 13 March 2026  00:47:46 +0000 (0:00:01.951)       0:00:12.235 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Friday 13 March 2026  00:47:49 +0000 (0:00:02.440)       0:00:14.675 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Friday 13 March 2026  00:47:49 +0000 (0:00:00.584)       0:00:15.259 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Friday 13 March 2026  00:47:50 +0000 (0:00:00.749)       0:00:16.009 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Friday 13 March 2026  00:47:51 +0000 (0:00:00.626)       0:00:16.635 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-13 00:47:47.390160', 'end': '2026-03-13 00:47:47.452558', 'delta': '0:00:00.062398', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-13 00:47:48.417953', 'end': '2026-03-13 00:47:48.478328', 'delta': '0:00:00.060375', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-13 00:47:48.949771', 'end': '2026-03-13 00:47:49.021265', 'delta': '0:00:00.071494', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Friday 13 March 2026  00:47:51 +0000 (0:00:00.297)       0:00:16.933 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Friday 13 March 2026  00:47:53 +0000 (0:00:01.855)       0:00:18.788 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Friday 13 March 2026  00:47:53 +0000 (0:00:00.576)       0:00:19.364 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Friday 13 March 2026  00:47:54 +0000 (0:00:01.095)       0:00:20.459 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Friday 13 March 2026  00:47:56 +0000 (0:00:01.814)       0:00:22.274 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Friday 13 March 2026  00:47:58 +0000 (0:00:01.456)       0:00:23.731 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Friday 13 March 2026  00:47:58 +0000 (0:00:00.266)       0:00:23.997 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Friday 13 March 2026  00:47:58 +0000 (0:00:00.514)       0:00:24.512 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Friday 13 March 2026  00:48:00 +0000 (0:00:01.542)       0:00:26.055 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Friday 13 March 2026  00:48:02 +0000 (0:00:02.334)       0:00:28.389 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Friday 13 March 2026  00:48:04 +0000 (0:00:01.569)       0:00:29.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Friday 13 March 2026  00:48:06 +0000 (0:00:01.652)       0:00:31.611 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Friday 13 March 2026  00:48:07 +0000 (0:00:01.369)       0:00:32.982 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Friday 13 March 2026  00:48:08 +0000 (0:00:01.261)       0:00:34.243 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Friday 13 March 2026  00:48:09 +0000 (0:00:00.687)       0:00:34.930 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-39-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Friday 13 March 2026  00:48:10 +0000 (0:00:01.447)       0:00:36.377 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16', 'scsi-SQEMU_QEMU_HARDDISK_dd0e771c-a216-451f-be9a-7a9338004557-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-39-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16', 'scsi-SQEMU_QEMU_HARDDISK_f3b73d97-a549-411a-96e1-9b740e8b4b02-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16', 'scsi-SQEMU_QEMU_HARDDISK_180093e4-fb06-417b-b075-79800beac41f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Friday 13 March 2026  00:48:11 +0000 (0:00:00.849)       0:00:37.227 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Friday 13 March 2026  00:48:12 +0000 (0:00:01.177)       0:00:38.404 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Friday 13 March 2026  00:48:13 +0000 (0:00:00.622)       0:00:39.027 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Friday 13 March 2026  00:48:14 +0000 (0:00:00.760)       0:00:39.787 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Friday 13 March 2026  00:48:15 +0000 (0:00:00.778)       0:00:40.566 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Friday 13 March 2026  00:48:16 +0000 (0:00:01.200)       0:00:41.766 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Friday 13 March 2026  00:48:17 +0000 (0:00:01.089)       0:00:42.856 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Friday 13 March 2026  00:48:20 +0000 (0:00:02.819)       0:00:45.676 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Friday 13 March 2026  00:48:21 +0000 (0:00:01.066)       0:00:46.742 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Friday 13 March 2026  00:48:22 +0000 (0:00:01.144)       0:00:47.887 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Friday 13 March 2026  00:48:22 +0000 (0:00:00.398)       0:00:48.286 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Friday 13 March 2026  00:48:23 +0000 (0:00:00.295)       0:00:48.582 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Friday 13 March 2026  00:48:23 +0000 (0:00:00.416)       0:00:48.999 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Friday 13 March 2026  00:48:23 +0000 (0:00:00.376)       0:00:49.375 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:00.355)       0:00:49.731 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:00.342)       0:00:50.073 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Friday 13 March 2026  00:48:24 +0000 (0:00:00.371)       0:00:50.445 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Friday 13 March 2026  00:48:25 +0000 (0:00:00.285)       0:00:50.730 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Friday 13 March 2026  00:48:25 +0000 (0:00:00.688)       0:00:51.418 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Friday 13 March 2026  00:48:26 +0000 (0:00:00.750)       0:00:52.169 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:48:28 +0000 (0:00:01.640)       0:00:53.809 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:48:29 +0000 (0:00:01.091)       0:00:54.900 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:48:30 +0000 (0:00:01.330)       0:00:56.231 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:48:32 +0000 (0:00:01.302)       0:00:57.533 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:48:32 +0000 (0:00:00.858)       0:00:58.392 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:48:33 +0000 (0:00:00.758)       0:00:59.150 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:48:34 +0000 (0:00:00.853)       0:01:00.004 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:48:36 +0000 (0:00:01.611)       0:01:01.615 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:48:37 +0000 (0:00:00.931)       0:01:02.547 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:48:37 +0000 (0:00:00.905)       0:01:03.453 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:48:39 +0000 (0:00:01.131)       0:01:04.584 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:48:40 +0000 (0:00:01.541)       0:01:06.126 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:48:41 +0000 (0:00:00.946)       0:01:07.072 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:48:42 +0000 (0:00:01.180)       0:01:08.253 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:48:44 +0000 (0:00:01.336)       0:01:09.590 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:48:45 +0000 (0:00:01.338)       0:01:10.928 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:48:46 +0000 (0:00:00.897)       0:01:11.826 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:48:47 +0000 (0:00:01.274)       0:01:13.101 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:48:48 +0000 (0:00:00.796)       0:01:13.898 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:48:49 +0000 (0:00:00.842)       0:01:14.741 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:48:50 +0000 (0:00:01.194)       0:01:15.935 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Friday 13 March 2026  00:48:52 +0000 (0:00:02.493)       0:01:18.429 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Friday 13 March 2026  00:48:55 +0000 (0:00:02.618)       0:01:21.047 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Friday 13 March 2026  00:48:59 +0000 (0:00:03.483)       0:01:24.531 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Friday 13 March 2026  00:49:00 +0000 (0:00:01.174)       0:01:25.705 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Friday 13 March 2026  00:49:01 +0000 (0:00:00.923)       0:01:26.628 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Friday 13 March 2026  00:49:02 +0000 (0:00:00.937)       0:01:27.566 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Friday 13 March 2026  00:49:03 +0000 (0:00:01.475)       0:01:29.041 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Friday 13 March 2026  00:49:04 +0000 (0:00:01.431)       0:01:30.472 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Friday 13 March 2026  00:49:05 +0000 (0:00:00.611)       0:01:31.084 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Friday 13 March 2026  00:49:06 +0000 (0:00:00.803)       0:01:31.888 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Friday 13 March 2026  00:49:06 +0000 (0:00:00.584)       0:01:32.472 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Friday 13 March 2026  00:49:08 +0000 (0:00:01.221)       0:01:33.694 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:42.045)       0:02:15.739 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Friday 13 March 2026  00:49:50 +0000 (0:00:00.580)       0:02:16.319 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Friday 13 March 2026  00:49:51 +0000 (0:00:00.847)       0:02:17.166 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Friday 13 March 2026  00:49:51 +0000 (0:00:00.161)       0:02:17.328 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Friday 13 March 2026  00:49:52 +0000 (0:00:00.580)       0:02:17.908 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Friday 13 March 2026  00:49:53 +0000 (0:00:00.658)       0:02:18.566 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Friday 13 March 2026  00:49:53 +0000 (0:00:00.721)       0:02:19.288 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Friday 13 March 2026  00:49:57 +0000 (0:00:03.520)       0:02:22.809 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Friday 13 March 2026  00:49:57 +0000 (0:00:00.605)       0:02:23.415 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Friday 13 March 2026  00:49:58 +0000 (0:00:01.034)       0:02:24.449 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Friday 13 March 2026  00:49:59 +0000 (0:00:00.654)       0:02:25.104 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Friday 13 March 2026  00:50:00 +0000 (0:00:00.497)       0:02:25.602 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Friday 13 March 2026  00:50:00 +0000 (0:00:00.656)       0:02:26.258 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Friday 13 March 2026  00:50:01 +0000 (0:00:00.520)       0:02:26.779 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Friday 13 March 2026  00:50:01 +0000 (0:00:00.639)       0:02:27.418 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Friday 13 March 2026  00:50:02 +0000 (0:00:00.536)       0:02:27.954 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Friday 13 March 2026  00:50:03 +0000 (0:00:00.799)       0:02:28.754 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Friday 13 March 2026  00:50:03 +0000 (0:00:00.612)       0:02:29.366 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Friday 13 March 2026  00:50:05 +0000 (0:00:01.256)       0:02:30.623 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Friday 13 March 2026  00:50:06 +0000 (0:00:01.327)       0:02:31.950 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Friday 13 March 2026  00:50:12 +0000 (0:00:06.457)       0:02:38.407 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Friday 13 March 2026  00:50:13 +0000 (0:00:00.878)       0:02:39.286 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Friday 13 March 2026  00:50:14 +0000 (0:00:00.918)       0:02:40.205 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Friday 13 March 2026  00:50:16 +0000 (0:00:01.629)       0:02:41.834 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Friday 13 March 2026  00:50:16 +0000 (0:00:00.681)       0:02:42.515 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Friday 13 March 2026  00:50:17 +0000 (0:00:00.903)       0:02:43.419 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Friday 13 March 2026  00:50:18 +0000 (0:00:00.565)       0:02:43.985 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Friday 13 March 2026  00:50:19 +0000 (0:00:00.741)       0:02:44.726 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Friday 13 March 2026  00:50:19 +0000 (0:00:00.585)       0:02:45.312 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Friday 13 March 2026  00:50:20 +0000 (0:00:00.782)       0:02:46.095 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Friday 13 March 2026  00:50:21 +0000 (0:00:00.797)       0:02:46.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Friday 13 March 2026  00:50:22 +0000 (0:00:01.065)       0:02:47.958 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Friday 13 March 2026  00:50:25 +0000 (0:00:02.929)       0:02:50.887 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Friday 13 March 2026  00:50:26 +0000 (0:00:00.879)       0:02:51.767 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Friday 13 March 2026  00:50:26 +0000 (0:00:00.739)       0:02:52.507 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Friday 13 March 2026  00:50:27 +0000 (0:00:00.748)       0:02:53.255 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Friday 13 March 2026  00:50:28 +0000 (0:00:00.728)       0:02:53.984 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Friday 13 March 2026  00:50:29 +0000 (0:00:01.233)       0:02:55.218 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:00.438)       0:02:55.656 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Friday 13 March 2026  00:50:30 +0000 (0:00:00.612)       0:02:56.269 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Friday 13 March 2026  00:50:31 +0000 (0:00:00.579)       0:02:56.848 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Friday 13 March 2026  00:50:32 +0000 (0:00:00.715)       0:02:57.564 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Friday 13 March 2026  00:50:32 +0000 (0:00:00.723)       0:02:58.288 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Friday 13 March 2026  00:50:34 +0000 (0:00:01.374)       0:02:59.662 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Friday 13 March 2026  00:50:34 +0000 (0:00:00.452)       0:03:00.114 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Friday 13 March 2026  00:50:34 +0000 (0:00:00.352)       0:03:00.467 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.339)       0:03:00.807 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Friday 13 March 2026  00:50:35 +0000 (0:00:00.562)       0:03:01.370 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Friday 13 March 2026  00:50:37 +0000 (0:00:01.808)       0:03:03.178 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:50:40 +0000 (0:00:02.425)       0:03:05.603 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Friday 13 March 2026  00:50:41 +0000 (0:00:01.377)       0:03:06.981 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Friday 13 March 2026  00:50:42 +0000 (0:00:00.868)       0:03:07.849 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Friday 13 March 2026  00:50:42 +0000 (0:00:00.233)       0:03:08.082 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Friday 13 March 2026  00:50:43 +0000 (0:00:01.197)       0:03:09.280 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Friday 13 March 2026  00:50:44 +0000 (0:00:00.574)       0:03:09.854 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Friday 13 March 2026  00:50:44 +0000 (0:00:00.346)       0:03:10.201 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Friday 13 March 2026  00:50:45 +0000 (0:00:00.984)       0:03:11.185 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Friday 13 March 2026  00:50:45 +0000 (0:00:00.322)       0:03:11.507 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:00.291)       0:03:11.799 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:00.214)       0:03:12.013 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:00.315)       0:03:12.329 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Friday 13 March 2026  00:50:46 +0000 (0:00:00.180)       0:03:12.510 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Friday 13 March 2026  00:50:47 +0000 (0:00:00.175)       0:03:12.685 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Friday 13 March 2026  00:50:47 +0000 (0:00:00.100)       0:03:12.786 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Friday 13 March 2026  00:50:47 +0000 (0:00:00.585)       0:03:13.371 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Friday 13 March 2026  00:50:48 +0000 (0:00:00.244)       0:03:13.616 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Friday 13 March 2026  00:50:48 +0000 (0:00:00.363)       0:03:13.979 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Friday 13 March 2026  00:50:48 +0000 (0:00:00.278)       0:03:14.258 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Friday 13 March 2026  00:50:48 +0000 (0:00:00.184)       0:03:14.443 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Friday 13 March 2026  00:50:49 +0000 (0:00:00.229)       0:03:14.673 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Friday 13 March 2026  00:50:50 +0000 (0:00:00.854)       0:03:15.528 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Friday 13 March 2026  00:50:50 +0000 (0:00:00.278)       0:03:15.806 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Friday 13 March 2026  00:50:51 +0000 (0:00:01.206)       0:03:17.012 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Friday 13 March 2026  00:50:52 +0000 (0:00:00.841)       0:03:17.854 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Friday 13 March 2026  00:50:52 +0000 (0:00:00.589)       0:03:18.444 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Friday 13 March 2026  00:50:53 +0000 (0:00:00.870)       0:03:19.314 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Friday 13 March 2026  00:50:54 +0000 (0:00:00.532)       0:03:19.847 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Friday 13 March 2026  00:50:55 +0000 (0:00:01.149)       0:03:20.997 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Friday 13 March 2026  00:50:56 +0000 (0:00:00.598)       0:03:21.595 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Friday 13 March 2026  00:50:56 +0000 (0:00:00.327)       0:03:21.923 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Friday 13 March 2026  00:50:57 +0000 (0:00:00.880)       0:03:22.803 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Friday 13 March 2026  00:50:58 +0000 (0:00:00.828)       0:03:23.632 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Friday 13 March 2026  00:50:58 +0000 (0:00:00.562)       0:03:24.194 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Friday 13 March 2026  00:50:59 +0000 (0:00:01.131)       0:03:25.326 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Friday 13 March 2026  00:51:00 +0000 (0:00:00.636)       0:03:25.963 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mon] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:51:01 +0000 (0:00:00.601)       0:03:26.564 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:51:01 +0000 (0:00:00.781)       0:03:27.346 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:51:02 +0000 (0:00:00.586)       0:03:27.932 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:51:03 +0000 (0:00:01.090)       0:03:29.023 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:51:03 +0000 (0:00:00.300)       0:03:29.323 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:51:04 +0000 (0:00:00.296)       0:03:29.620 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:51:04 +0000 (0:00:00.239)       0:03:29.860 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:51:05 +0000 (0:00:00.821)       0:03:30.681 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:51:05 +0000 (0:00:00.256)       0:03:30.938 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:51:05 +0000 (0:00:00.304)       0:03:31.243 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:51:06 +0000 (0:00:00.730)       0:03:31.974 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:51:07 +0000 (0:00:00.769)       0:03:32.743 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:51:07 +0000 (0:00:00.267)       0:03:33.010 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:51:07 +0000 (0:00:00.287)       0:03:33.298 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:51:08 +0000 (0:00:00.283)       0:03:33.581 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:51:08 +0000 (0:00:00.250)       0:03:33.832 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:51:08 +0000 (0:00:00.439)       0:03:34.272 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:51:09 +0000 (0:00:00.266)       0:03:34.539 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:51:09 +0000 (0:00:00.254)       0:03:34.793 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:51:09 +0000 (0:00:00.264)       0:03:35.058 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:51:09 +0000 (0:00:00.415)       0:03:35.473 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Friday 13 March 2026  00:51:10 +0000 (0:00:00.499)       0:03:35.973 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Friday 13 March 2026  00:51:10 +0000 (0:00:00.284)       0:03:36.258 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Friday 13 March 2026  00:51:11 +0000 (0:00:00.708)       0:03:36.967 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Friday 13 March 2026  00:51:11 +0000 (0:00:00.147)       0:03:37.114 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Friday 13 March 2026  00:51:12 +0000 (0:00:01.096)       0:03:38.210 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Friday 13 March 2026  00:51:13 +0000 (0:00:00.336)       0:03:38.547 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Friday 13 March 2026  00:51:13 +0000 (0:00:00.545)       0:03:39.092 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Friday 13 March 2026  00:51:14 +0000 (0:00:01.163)       0:03:40.256 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Friday 13 March 2026  00:51:15 +0000 (0:00:00.790)       0:03:41.046 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Friday 13 March 2026  00:51:16 +0000 (0:00:00.695)       0:03:41.742 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Friday 13 March 2026  00:51:16 +0000 (0:00:00.635)       0:03:42.377 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Friday 13 March 2026  00:51:18 +0000 (0:00:01.575)       0:03:43.952 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Friday 13 March 2026  00:51:19 +0000 (0:00:00.706)       0:03:44.659 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Friday 13 March 2026  00:51:22 +0000 (0:00:03.007)       0:03:47.666 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Friday 13 March 2026  00:51:23 +0000 (0:00:01.304)       0:03:48.971 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Friday 13 March 2026  00:51:23 +0000 (0:00:00.339)       0:03:49.310 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Friday 13 March 2026  00:51:24 +0000 (0:00:00.507)       0:03:49.818 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Friday 13 March 2026  00:51:25 +0000 (0:00:01.534)       0:03:51.352 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Friday 13 March 2026  00:51:27 +0000 (0:00:01.209)       0:03:52.561 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Friday 13 March 2026  00:51:27 +0000 (0:00:00.218)       0:03:52.780 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Friday 13 March 2026  00:51:27 +0000 (0:00:00.594)       0:03:53.374 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Friday 13 March 2026  00:51:28 +0000 (0:00:00.284)       0:03:53.659 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Friday 13 March 2026  00:51:28 +0000 (0:00:00.255)       0:03:53.914 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Friday 13 March 2026  00:51:29 +0000 (0:00:00.638)       0:03:54.553 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Friday 13 March 2026  00:51:31 +0000 (0:00:01.980)       0:03:56.534 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Friday 13 March 2026  00:51:32 +0000 (0:00:01.205)       0:03:57.739 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Friday 13 March 2026  00:51:33 +0000 (0:00:01.737)       0:03:59.477 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Friday 13 March 2026  00:51:36 +0000 (0:00:02.648)       0:04:02.125 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Friday 13 March 2026  00:51:37 +0000 (0:00:00.542)       0:04:02.668 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Friday 13 March 2026  00:51:58 +0000 (0:00:21.791)       0:04:24.459 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Friday 13 March 2026  00:52:07 +0000 (0:00:08.523)       0:04:32.983 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Friday 13 March 2026  00:52:08 +0000 (0:00:00.567)       0:04:33.551 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__ae91f54b06fcc83089cdafa34f2079a7ae738feb'}]) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:52:22 +0000 (0:00:14.641)       0:04:48.193 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Friday 13 March 2026  00:52:22 +0000 (0:00:00.314)       0:04:48.508 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Friday 13 March 2026  00:52:23 +0000 (0:00:00.816)       0:04:49.324 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Friday 13 March 2026  00:52:24 +0000 (0:00:00.316)       0:04:49.640 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Friday 13 March 2026  00:52:24 +0000 (0:00:00.330)       0:04:49.971 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Friday 13 March 2026  00:52:25 +0000 (0:00:00.870)       0:04:50.842 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mgr] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:52:26 +0000 (0:00:00.793)       0:04:51.636 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:52:26 +0000 (0:00:00.488)       0:04:52.124 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:52:27 +0000 (0:00:00.750)       0:04:52.874 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:52:28 +0000 (0:00:00.754)       0:04:53.629 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:52:28 +0000 (0:00:00.270)       0:04:53.899 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:52:28 +0000 (0:00:00.426)       0:04:54.325 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:52:29 +0000 (0:00:00.338)       0:04:54.664 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:52:29 +0000 (0:00:00.682)       0:04:55.346 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:52:30 +0000 (0:00:00.263)       0:04:55.609 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:52:30 +0000 (0:00:00.420)       0:04:56.029 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:52:31 +0000 (0:00:00.705)       0:04:56.734 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:52:31 +0000 (0:00:00.648)       0:04:57.383 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:00.234)       0:04:57.617 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:00.475)       0:04:58.093 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:52:32 +0000 (0:00:00.243)       0:04:58.337 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:52:33 +0000 (0:00:00.270)       0:04:58.607 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:52:33 +0000 (0:00:00.357)       0:04:58.965 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:52:33 +0000 (0:00:00.286)       0:04:59.252 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:52:34 +0000 (0:00:00.550)       0:04:59.802 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:52:34 +0000 (0:00:00.359)       0:05:00.161 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:52:34 +0000 (0:00:00.289)       0:05:00.451 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Friday 13 March 2026  00:52:35 +0000 (0:00:00.589)       0:05:01.040 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.542)       0:05:01.583 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Friday 13 March 2026  00:52:36 +0000 (0:00:00.430)       0:05:02.013 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Friday 13 March 2026  00:52:37 +0000 (0:00:00.594)       0:05:02.608 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Friday 13 March 2026  00:52:37 +0000 (0:00:00.408)       0:05:03.016 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Friday 13 March 2026  00:52:47 +0000 (0:00:09.626)       0:05:12.642 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Friday 13 March 2026  00:52:47 +0000 (0:00:00.285)       0:05:12.928 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Friday 13 March 2026  00:52:49 +0000 (0:00:01.914)       0:05:14.843 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Friday 13 March 2026  00:52:50 +0000 (0:00:01.502)       0:05:16.346 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Friday 13 March 2026  00:52:51 +0000 (0:00:00.779)       0:05:17.125 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Friday 13 March 2026  00:52:51 +0000 (0:00:00.291)       0:05:17.417 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Friday 13 March 2026  00:52:52 +0000 (0:00:00.268)       0:05:17.685 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Friday 13 March 2026  00:52:52 +0000 (0:00:00.568)       0:05:18.254 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Friday 13 March 2026  00:52:53 +0000 (0:00:00.322)       0:05:18.577 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Friday 13 March 2026  00:52:53 +0000 (0:00:00.267)       0:05:18.844 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Friday 13 March 2026  00:52:53 +0000 (0:00:00.496)       0:05:19.340 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Friday 13 March 2026  00:52:54 +0000 (0:00:01.064)       0:05:20.405 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Friday 13 March 2026  00:52:56 +0000 (0:00:01.182)       0:05:21.587 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Friday 13 March 2026  00:52:58 +0000 (0:00:01.967)       0:05:23.555 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Friday 13 March 2026  00:53:00 +0000 (0:00:02.572)       0:05:26.127 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Friday 13 March 2026  00:53:01 +0000 (0:00:00.423)       0:05:26.551 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Friday 13 March 2026  00:53:37 +0000 (0:00:36.006)       0:06:02.557 ********** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Friday 13 March 2026  00:53:38 +0000 (0:00:01.368)       0:06:03.926 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Friday 13 March 2026  00:53:38 +0000 (0:00:00.304)       0:06:04.231 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Friday 13 March 2026  00:53:38 +0000 (0:00:00.107)       0:06:04.338 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Friday 13 March 2026  00:53:45 +0000 (0:00:06.638)       0:06:10.976 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:53:50 +0000 (0:00:04.679)       0:06:15.656 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Friday 13 March 2026  00:53:50 +0000 (0:00:00.593)       0:06:16.250 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Friday 13 March 2026  00:53:51 +0000 (0:00:00.576)       0:06:16.827 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Friday 13 March 2026  00:53:51 +0000 (0:00:00.266)       0:06:17.093 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Friday 13 March 2026  00:53:52 +0000 (0:00:00.996)       0:06:18.090 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Friday 13 March 2026  00:53:53 +0000 (0:00:00.525)       0:06:18.616 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-osd] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:53:53 +0000 (0:00:00.663)       0:06:19.279 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:53:54 +0000 (0:00:00.441)       0:06:19.721 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:53:54 +0000 (0:00:00.592)       0:06:20.313 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:53:55 +0000 (0:00:00.258)       0:06:20.572 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:53:55 +0000 (0:00:00.670)       0:06:21.243 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:53:56 +0000 (0:00:00.670)       0:06:21.913 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:53:57 +0000 (0:00:01.022)       0:06:22.936 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:53:57 +0000 (0:00:00.257)       0:06:23.194 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:53:57 +0000 (0:00:00.320)       0:06:23.514 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:53:58 +0000 (0:00:00.223)       0:06:23.738 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:53:59 +0000 (0:00:00.817)       0:06:24.556 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:53:59 +0000 (0:00:00.666)       0:06:25.223 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:53:59 +0000 (0:00:00.208)       0:06:25.432 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:54:00 +0000 (0:00:00.263)       0:06:25.696 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:54:00 +0000 (0:00:00.437)       0:06:26.133 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:54:00 +0000 (0:00:00.280)       0:06:26.414 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:54:01 +0000 (0:00:00.296)       0:06:26.710 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:54:01 +0000 (0:00:00.249)       0:06:26.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:54:01 +0000 (0:00:00.409)       0:06:27.368 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:54:02 +0000 (0:00:00.251)       0:06:27.620 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:54:02 +0000 (0:00:00.276)       0:06:27.896 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Friday 13 March 2026  00:54:02 +0000 (0:00:00.562)       0:06:28.458 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Friday 13 March 2026  00:54:03 +0000 (0:00:00.279)       0:06:28.738 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Friday 13 March 2026  00:54:03 +0000 (0:00:00.548)       0:06:29.286 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Friday 13 March 2026  00:54:04 +0000 (0:00:00.438)       0:06:29.724 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Friday 13 March 2026  00:54:04 +0000 (0:00:00.399)       0:06:30.124 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Friday 13 March 2026  00:54:04 +0000 (0:00:00.263)       0:06:30.387 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Friday 13 March 2026  00:54:05 +0000 (0:00:00.663)       0:06:31.051 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Friday 13 March 2026  00:54:05 +0000 (0:00:00.429)       0:06:31.481 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Friday 13 March 2026  00:54:09 +0000 (0:00:03.744)       0:06:35.226 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Friday 13 March 2026  00:54:10 +0000 (0:00:00.305)       0:06:35.532 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Friday 13 March 2026  00:54:10 +0000 (0:00:00.403)       0:06:35.936 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Friday 13 March 2026  00:54:11 +0000 (0:00:01.248)       0:06:37.184 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Friday 13 March 2026  00:54:13 +0000 (0:00:02.244)       0:06:39.428 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Friday 13 March 2026  00:54:15 +0000 (0:00:01.282)       0:06:40.711 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Friday 13 March 2026  00:54:17 +0000 (0:00:02.626)       0:06:43.337 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Use ceph-volume to create osds] *******************************",
                                "Friday 13 March 2026  00:54:18 +0000 (0:00:00.746)       0:06:44.084 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-14d46521-cd63-5c40-b949-080247a26419', 'data_vg': 'ceph-14d46521-cd63-5c40-b949-080247a26419'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-34f0c902-d036-5088-bfac-53928dad7c62', 'data_vg': 'ceph-34f0c902-d036-5088-bfac-53928dad7c62'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-a323501f-4860-533e-be9d-aa017f55891f', 'data_vg': 'ceph-a323501f-4860-533e-be9d-aa017f55891f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-74a7d81d-0a8f-5a99-882a-be09baf5f158', 'data_vg': 'ceph-74a7d81d-0a8f-5a99-882a-be09baf5f158'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-02d8354f-c2be-5171-944a-5a76777090d4', 'data_vg': 'ceph-02d8354f-c2be-5171-944a-5a76777090d4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-27536765-3469-5fa1-a30d-2781c44211ac', 'data_vg': 'ceph-27536765-3469-5fa1-a30d-2781c44211ac'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Friday 13 March 2026  00:55:04 +0000 (0:00:45.469)       0:07:29.554 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Friday 13 March 2026  00:55:04 +0000 (0:00:00.378)       0:07:29.932 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-4, testbed-node-5, testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Friday 13 March 2026  00:55:05 +0000 (0:00:00.839)       0:07:30.772 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Friday 13 March 2026  00:55:06 +0000 (0:00:00.802)       0:07:31.574 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Friday 13 March 2026  00:55:09 +0000 (0:00:03.032)       0:07:34.606 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Friday 13 March 2026  00:55:09 +0000 (0:00:00.707)       0:07:35.314 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Friday 13 March 2026  00:55:10 +0000 (0:00:01.188)       0:07:36.503 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Friday 13 March 2026  00:55:12 +0000 (0:00:01.109)       0:07:37.612 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Friday 13 March 2026  00:55:14 +0000 (0:00:01.927)       0:07:39.539 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Friday 13 March 2026  00:55:14 +0000 (0:00:00.580)       0:07:40.120 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Friday 13 March 2026  00:55:14 +0000 (0:00:00.320)       0:07:40.441 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Friday 13 March 2026  00:55:16 +0000 (0:00:01.284)       0:07:41.725 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Friday 13 March 2026  00:55:18 +0000 (0:00:02.365)       0:07:44.091 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Friday 13 March 2026  00:55:22 +0000 (0:00:03.566)       0:07:47.657 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Friday 13 March 2026  00:55:25 +0000 (0:00:02.996)       0:07:50.654 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Friday 13 March 2026  00:55:37 +0000 (0:00:12.509)       0:08:03.164 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:55:38 +0000 (0:00:00.994)       0:08:04.159 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Friday 13 March 2026  00:55:38 +0000 (0:00:00.340)       0:08:04.499 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Friday 13 March 2026  00:55:39 +0000 (0:00:00.459)       0:08:04.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Friday 13 March 2026  00:55:40 +0000 (0:00:00.656)       0:08:05.615 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Friday 13 March 2026  00:55:40 +0000 (0:00:00.295)       0:08:05.911 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Friday 13 March 2026  00:55:40 +0000 (0:00:00.217)       0:08:06.129 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Friday 13 March 2026  00:55:40 +0000 (0:00:00.262)       0:08:06.391 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Friday 13 March 2026  00:55:41 +0000 (0:00:00.192)       0:08:06.583 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Friday 13 March 2026  00:55:41 +0000 (0:00:00.198)       0:08:06.782 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Friday 13 March 2026  00:55:41 +0000 (0:00:00.111)       0:08:06.894 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Friday 13 March 2026  00:55:41 +0000 (0:00:00.180)       0:08:07.075 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Friday 13 March 2026  00:55:42 +0000 (0:00:00.561)       0:08:07.637 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Friday 13 March 2026  00:55:42 +0000 (0:00:00.341)       0:08:07.978 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Friday 13 March 2026  00:55:42 +0000 (0:00:00.311)       0:08:08.289 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Friday 13 March 2026  00:55:42 +0000 (0:00:00.183)       0:08:08.473 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-crash] ***************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:55:43 +0000 (0:00:00.720)       0:08:09.194 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:55:44 +0000 (0:00:00.941)       0:08:10.135 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:55:45 +0000 (0:00:00.877)       0:08:11.013 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:55:46 +0000 (0:00:01.054)       0:08:12.067 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:55:47 +0000 (0:00:00.668)       0:08:12.736 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:55:48 +0000 (0:00:00.802)       0:08:13.539 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:55:48 +0000 (0:00:00.635)       0:08:14.174 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:55:49 +0000 (0:00:01.003)       0:08:15.178 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:55:50 +0000 (0:00:00.510)       0:08:15.688 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:55:50 +0000 (0:00:00.690)       0:08:16.379 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:55:51 +0000 (0:00:00.991)       0:08:17.371 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:55:53 +0000 (0:00:01.383)       0:08:18.755 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:55:53 +0000 (0:00:00.615)       0:08:19.370 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:55:54 +0000 (0:00:00.860)       0:08:20.231 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:55:55 +0000 (0:00:00.570)       0:08:20.801 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:55:56 +0000 (0:00:00.818)       0:08:21.620 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:55:56 +0000 (0:00:00.579)       0:08:22.200 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:55:57 +0000 (0:00:00.777)       0:08:22.977 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:55:58 +0000 (0:00:00.569)       0:08:23.547 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:55:58 +0000 (0:00:00.805)       0:08:24.353 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:55:59 +0000 (0:00:00.596)       0:08:24.949 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Friday 13 March 2026  00:56:00 +0000 (0:00:01.222)       0:08:26.171 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Friday 13 March 2026  00:56:04 +0000 (0:00:03.585)       0:08:29.756 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Friday 13 March 2026  00:56:06 +0000 (0:00:02.244)       0:08:32.001 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Friday 13 March 2026  00:56:08 +0000 (0:00:01.887)       0:08:33.888 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Friday 13 March 2026  00:56:09 +0000 (0:00:00.940)       0:08:34.829 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Friday 13 March 2026  00:56:10 +0000 (0:00:01.245)       0:08:36.074 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Friday 13 March 2026  00:56:12 +0000 (0:00:01.633)       0:08:37.708 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************",
                                "Friday 13 March 2026  00:56:15 +0000 (0:00:03.046)       0:08:40.754 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******",
                                "Friday 13 March 2026  00:56:16 +0000 (0:00:01.299)       0:08:42.053 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************",
                                "Friday 13 March 2026  00:56:17 +0000 (0:00:00.824)       0:08:42.878 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******",
                                "Friday 13 March 2026  00:56:19 +0000 (0:00:02.209)       0:08:45.087 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mds] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.920)       0:08:46.008 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.454)       0:08:46.462 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:56:21 +0000 (0:00:00.739)       0:08:47.202 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:56:21 +0000 (0:00:00.254)       0:08:47.456 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:56:22 +0000 (0:00:00.634)       0:08:48.091 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:56:23 +0000 (0:00:00.806)       0:08:48.897 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:00.741)       0:08:49.639 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:00.306)       0:08:49.945 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:00.291)       0:08:50.236 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:56:25 +0000 (0:00:00.585)       0:08:50.822 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:56:26 +0000 (0:00:00.806)       0:08:51.629 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:56:26 +0000 (0:00:00.775)       0:08:52.404 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:56:27 +0000 (0:00:00.314)       0:08:52.719 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:56:27 +0000 (0:00:00.575)       0:08:53.294 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:56:28 +0000 (0:00:00.326)       0:08:53.622 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:56:28 +0000 (0:00:00.361)       0:08:53.983 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:56:28 +0000 (0:00:00.343)       0:08:54.326 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:56:29 +0000 (0:00:00.623)       0:08:54.950 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:56:29 +0000 (0:00:00.308)       0:08:55.259 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:56:30 +0000 (0:00:00.309)       0:08:55.568 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:56:30 +0000 (0:00:00.321)       0:08:55.890 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Friday 13 March 2026  00:56:31 +0000 (0:00:00.776)       0:08:56.666 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule details] *********************",
                                "Friday 13 March 2026  00:56:31 +0000 (0:00:00.426)       0:08:57.093 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule name] ************************",
                                "Friday 13 March 2026  00:56:33 +0000 (0:00:01.888)       0:08:58.981 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create filesystem pools] **************************************",
                                "Friday 13 March 2026  00:56:33 +0000 (0:00:00.242)       0:08:59.224 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create ceph filesystem] ***************************************",
                                "Friday 13 March 2026  00:56:42 +0000 (0:00:08.714)       0:09:07.938 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Friday 13 March 2026  00:56:46 +0000 (0:00:03.786)       0:09:11.725 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Friday 13 March 2026  00:56:46 +0000 (0:00:00.451)       0:09:12.177 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Friday 13 March 2026  00:56:47 +0000 (0:00:01.153)       0:09:13.331 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Friday 13 March 2026  00:56:50 +0000 (0:00:02.738)       0:09:16.069 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Friday 13 March 2026  00:56:51 +0000 (0:00:01.385)       0:09:17.455 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Friday 13 March 2026  00:56:54 +0000 (0:00:02.308)       0:09:19.764 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Friday 13 March 2026  00:56:54 +0000 (0:00:00.237)       0:09:20.002 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Friday 13 March 2026  00:56:55 +0000 (0:00:00.620)       0:09:20.623 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Friday 13 March 2026  00:56:55 +0000 (0:00:00.465)       0:09:21.088 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Friday 13 March 2026  00:56:56 +0000 (0:00:01.065)       0:09:22.154 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Friday 13 March 2026  00:56:58 +0000 (0:00:01.412)       0:09:23.566 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Friday 13 March 2026  00:56:59 +0000 (0:00:01.696)       0:09:25.263 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Friday 13 March 2026  00:57:01 +0000 (0:00:01.877)       0:09:27.141 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:57:02 +0000 (0:00:01.266)       0:09:28.407 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Friday 13 March 2026  00:57:03 +0000 (0:00:00.612)       0:09:29.020 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Friday 13 March 2026  00:57:04 +0000 (0:00:00.697)       0:09:29.718 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Friday 13 March 2026  00:57:04 +0000 (0:00:00.301)       0:09:30.019 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Friday 13 March 2026  00:57:05 +0000 (0:00:01.200)       0:09:31.220 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Friday 13 March 2026  00:57:06 +0000 (0:00:00.658)       0:09:31.879 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Friday 13 March 2026  00:57:06 +0000 (0:00:00.575)       0:09:32.455 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Friday 13 March 2026  00:57:07 +0000 (0:00:00.437)       0:09:32.892 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Friday 13 March 2026  00:57:07 +0000 (0:00:00.538)       0:09:33.431 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Friday 13 March 2026  00:57:08 +0000 (0:00:00.257)       0:09:33.688 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Friday 13 March 2026  00:57:09 +0000 (0:00:00.869)       0:09:34.557 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Friday 13 March 2026  00:57:09 +0000 (0:00:00.949)       0:09:35.507 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Friday 13 March 2026  00:57:10 +0000 (0:00:00.697)       0:09:36.204 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Friday 13 March 2026  00:57:10 +0000 (0:00:00.259)       0:09:36.463 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Friday 13 March 2026  00:57:11 +0000 (0:00:00.288)       0:09:36.752 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Friday 13 March 2026  00:57:11 +0000 (0:00:00.413)       0:09:37.166 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Friday 13 March 2026  00:57:12 +0000 (0:00:00.650)       0:09:37.817 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Friday 13 March 2026  00:57:12 +0000 (0:00:00.605)       0:09:38.422 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Friday 13 March 2026  00:57:13 +0000 (0:00:00.256)       0:09:38.679 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Friday 13 March 2026  00:57:13 +0000 (0:00:00.409)       0:09:39.088 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Friday 13 March 2026  00:57:13 +0000 (0:00:00.274)       0:09:39.362 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Friday 13 March 2026  00:57:14 +0000 (0:00:00.276)       0:09:39.639 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Friday 13 March 2026  00:57:14 +0000 (0:00:00.279)       0:09:39.919 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Friday 13 March 2026  00:57:14 +0000 (0:00:00.414)       0:09:40.333 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Friday 13 March 2026  00:57:15 +0000 (0:00:00.276)       0:09:40.609 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Friday 13 March 2026  00:57:15 +0000 (0:00:00.295)       0:09:40.905 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Friday 13 March 2026  00:57:15 +0000 (0:00:00.280)       0:09:41.185 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Friday 13 March 2026  00:57:16 +0000 (0:00:00.587)       0:09:41.772 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Friday 13 March 2026  00:57:16 +0000 (0:00:00.456)       0:09:42.229 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Friday 13 March 2026  00:57:18 +0000 (0:00:01.973)       0:09:44.202 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Friday 13 March 2026  00:57:20 +0000 (0:00:01.547)       0:09:45.749 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Friday 13 March 2026  00:57:20 +0000 (0:00:00.315)       0:09:46.065 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Friday 13 March 2026  00:57:21 +0000 (0:00:00.515)       0:09:46.580 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Friday 13 March 2026  00:57:22 +0000 (0:00:01.426)       0:09:48.007 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Friday 13 March 2026  00:57:26 +0000 (0:00:04.077)       0:09:52.084 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Friday 13 March 2026  00:57:28 +0000 (0:00:02.092)       0:09:54.177 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Friday 13 March 2026  00:57:30 +0000 (0:00:01.382)       0:09:55.559 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Friday 13 March 2026  00:57:30 +0000 (0:00:00.223)       0:09:55.782 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Friday 13 March 2026  00:57:31 +0000 (0:00:01.154)       0:09:56.937 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Friday 13 March 2026  00:57:32 +0000 (0:00:00.598)       0:09:57.535 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Friday 13 March 2026  00:58:02 +0000 (0:00:30.025)       0:10:27.561 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Friday 13 March 2026  00:58:02 +0000 (0:00:00.254)       0:10:27.816 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Friday 13 March 2026  00:58:02 +0000 (0:00:00.298)       0:10:28.114 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Friday 13 March 2026  00:58:03 +0000 (0:00:00.611)       0:10:28.725 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Friday 13 March 2026  00:58:03 +0000 (0:00:00.460)       0:10:29.186 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Friday 13 March 2026  00:58:04 +0000 (0:00:01.133)       0:10:30.319 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Friday 13 March 2026  00:58:06 +0000 (0:00:01.226)       0:10:31.546 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Friday 13 March 2026  00:58:07 +0000 (0:00:01.607)       0:10:33.154 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Friday 13 March 2026  00:58:10 +0000 (0:00:03.299)       0:10:36.454 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Friday 13 March 2026  00:58:11 +0000 (0:00:00.303)       0:10:36.758 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Friday 13 March 2026  00:58:11 +0000 (0:00:00.444)       0:10:37.202 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Friday 13 March 2026  00:58:12 +0000 (0:00:00.446)       0:10:37.649 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Friday 13 March 2026  00:58:12 +0000 (0:00:00.296)       0:10:37.945 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Friday 13 March 2026  00:58:12 +0000 (0:00:00.533)       0:10:38.479 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=32  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:58:13 +0000 (0:00:00.207)       0:10:38.687 ********** ",
                                "=============================================================================== ",
                                "ceph-osd : Use ceph-volume to create osds ------------------------------ 45.47s",
                                "ceph-container-common : Pulling Ceph container image ------------------- 42.05s",
                                "ceph-mgr : Wait for all mgr to be up ----------------------------------- 36.01s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 30.03s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.79s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 14.64s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 12.51s",
                                "ceph-mgr : Create ceph mgr keyring(s) on a mon node --------------------- 9.63s",
                                "ceph-mds : Create filesystem pools -------------------------------------- 8.71s",
                                "ceph-mon : Fetch ceph initial keys -------------------------------------- 8.52s",
                                "ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.64s",
                                "ceph-config : Create ceph initial directories --------------------------- 6.46s",
                                "ceph-mgr : Add modules to ceph-mgr -------------------------------------- 4.68s",
                                "ceph-rgw : Create rgw keyrings ------------------------------------------ 4.08s",
                                "ceph-mds : Create ceph filesystem --------------------------------------- 3.79s",
                                "ceph-osd : Apply operating system tuning -------------------------------- 3.75s",
                                "ceph-crash : Create client.crash keyring -------------------------------- 3.59s",
                                "ceph-osd : Systemd start osd -------------------------------------------- 3.57s",
                                "ceph-container-common : Get ceph version -------------------------------- 3.52s",
                                "ceph-container-common : Enable ceph.target ------------------------------ 3.48s",
                                "\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.248)       0:00:00.248 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.317)       0:00:00.565 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Friday 13 March 2026  00:56:21 +0000 (0:00:00.439)       0:00:01.005 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Friday 13 March 2026  00:56:21 +0000 (0:00:00.467)       0:00:01.472 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:02.577)       0:00:04.049 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Friday 13 March 2026  00:56:26 +0000 (0:00:02.173)       0:00:06.223 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Friday 13 March 2026  00:56:27 +0000 (0:00:00.946)       0:00:07.170 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:56:30 +0000 (0:00:03.011)       0:00:10.182 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Friday 13 March 2026  00:56:32 +0000 (0:00:01.529)       0:00:11.712 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Friday 13 March 2026  00:56:33 +0000 (0:00:01.285)       0:00:12.997 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Friday 13 March 2026  00:56:36 +0000 (0:00:02.631)       0:00:15.629 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Friday 13 March 2026  00:56:38 +0000 (0:00:02.498)       0:00:18.127 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : opensearch | Check containers] ****************",
                                "Friday 13 March 2026  00:56:40 +0000 (0:00:02.125)       0:00:20.253 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : opensearch | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:56:42 +0000 (0:00:01.844)       0:00:22.098 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:56:42 +0000 (0:00:00.450)       0:00:22.548 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2025.1', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal'], 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2025.1', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk GET /api/status']}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password', 'backend_http_extra': ['option httpchk  GET /api/status']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:01.233)       0:00:23.782 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:00.303)       0:00:24.085 ********** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:00.063)       0:00:24.148 ********** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:00.063)       0:00:24.212 ********** ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:00.065)       0:00:24.277 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Friday 13 March 2026  00:56:44 +0000 (0:00:00.195)       0:00:24.472 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Friday 13 March 2026  00:56:45 +0000 (0:00:00.217)       0:00:24.690 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Friday 13 March 2026  00:57:37 +0000 (0:00:52.408)       0:01:17.098 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Friday 13 March 2026  00:58:35 +0000 (0:00:57.632)       0:02:14.730 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Friday 13 March 2026  00:58:35 +0000 (0:00:00.543)       0:02:15.273 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch cluster to become healthy] **************",
                                "Friday 13 March 2026  00:58:37 +0000 (0:00:02.169)       0:02:17.443 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Friday 13 March 2026  00:58:39 +0000 (0:00:01.854)       0:02:19.297 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Friday 13 March 2026  00:58:42 +0000 (0:00:02.462)       0:02:21.760 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Friday 13 March 2026  00:58:44 +0000 (0:00:02.214)       0:02:23.974 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:58:46 +0000 (0:00:02.050)       0:02:26.024 ********** ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 57.63s",
                                "opensearch : Restart opensearch container ------------------------------ 52.41s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 3.01s",
                                "opensearch : Copying over config.json files for services ---------------- 2.63s",
                                "opensearch : Setting sysctl values -------------------------------------- 2.58s",
                                "opensearch : Copying over opensearch service config file ---------------- 2.50s",
                                "opensearch : Check if a log retention policy exists --------------------- 2.46s",
                                "opensearch : Create new log retention policy ---------------------------- 2.21s",
                                "opensearch : Ensuring config directories exist -------------------------- 2.17s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 2.17s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 2.13s",
                                "opensearch : Apply retention policy to existing indices ----------------- 2.05s",
                                "opensearch : Wait for OpenSearch cluster to become healthy -------------- 1.85s",
                                "service-check-containers : opensearch | Check containers ---------------- 1.84s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.53s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 1.29s",
                                "service-check-containers : Include tasks -------------------------------- 1.23s",
                                "opensearch : include_tasks ---------------------------------------------- 0.95s",
                                "opensearch : include_tasks ---------------------------------------------- 0.54s",
                                "opensearch : include_tasks ---------------------------------------------- 0.47s",
                                "\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fe920d26-8a98-41f8-af12-a178724756c1 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 511e09a0-cf07-498c-b6d5-d6f7cb4d260b is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Set kolla_action_mariadb] ************************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.102)       0:00:00.102 ********** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check MariaDB service] ***************************************************",
                                "Friday 13 March 2026  00:56:20 +0000 (0:00:00.030)       0:00:00.133 ********** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********",
                                "Friday 13 March 2026  00:56:23 +0000 (0:00:02.861)       0:00:02.994 ********** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = kolla_action_ng] ******************************",
                                "Friday 13 March 2026  00:56:23 +0000 (0:00:00.053)       0:00:03.048 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:56:23 +0000 (0:00:00.158)       0:00:03.207 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:00.324)       0:00:03.532 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Friday 13 March 2026  00:56:24 +0000 (0:00:00.606)       0:00:04.138 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:56:25 +0000 (0:00:00.390)       0:00:04.528 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Friday 13 March 2026  00:56:25 +0000 (0:00:00.618)       0:00:05.147 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Friday 13 March 2026  00:56:29 +0000 (0:00:03.402)       0:00:08.549 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Friday 13 March 2026  00:56:29 +0000 (0:00:00.640)       0:00:09.190 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Friday 13 March 2026  00:56:31 +0000 (0:00:01.488)       0:00:10.678 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Friday 13 March 2026  00:56:35 +0000 (0:00:03.842)       0:00:14.521 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Friday 13 March 2026  00:56:36 +0000 (0:00:01.124)       0:00:15.646 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:56:40 +0000 (0:00:04.183)       0:00:19.829 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  00:56:40 +0000 (0:00:00.544)       0:00:20.373 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:56:43 +0000 (0:00:02.661)       0:00:23.035 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  00:56:46 +0000 (0:00:02.524)       0:00:25.559 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : mariadb | Check containers] *******************",
                                "Friday 13 March 2026  00:56:48 +0000 (0:00:02.688)       0:00:28.248 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/l\u001b[32m2026-03-13 00:59:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "ocaltime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : mariadb | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:56:51 +0000 (0:00:02.508)       0:00:30.756 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:56:51 +0000 (0:00:00.561)       0:00:31.318 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Checking for mariadb cluster] **********************************",
                                "Friday 13 March 2026  00:56:54 +0000 (0:00:02.354)       0:00:33.672 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Cleaning up temp file on localhost] ****************************",
                                "Friday 13 March 2026  00:56:54 +0000 (0:00:00.331)       0:00:34.004 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Stop MariaDB containers] ***************************************",
                                "Friday 13 March 2026  00:56:54 +0000 (0:00:00.159)       0:00:34.163 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Run MariaDB wsrep recovery] ************************************",
                                "Friday 13 March 2026  00:56:55 +0000 (0:00:00.490)       0:00:34.653 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying MariaDB log file to /tmp] ******************************",
                                "Friday 13 March 2026  00:56:55 +0000 (0:00:00.365)       0:00:35.019 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB wsrep recovery seqno] ******************************",
                                "Friday 13 March 2026  00:56:55 +0000 (0:00:00.329)       0:00:35.348 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Removing MariaDB log file from /tmp] ***************************",
                                "Friday 13 March 2026  00:56:56 +0000 (0:00:00.284)       0:00:35.633 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Registering MariaDB seqno variable] ****************************",
                                "Friday 13 March 2026  00:56:56 +0000 (0:00:00.501)       0:00:36.134 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Comparing seqno value on all mariadb hosts] ********************",
                                "Friday 13 March 2026  00:56:56 +0000 (0:00:00.320)       0:00:36.454 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Writing hostname of host with the largest seqno to temp file] ***",
                                "Friday 13 March 2026  00:56:57 +0000 (0:00:00.321)       0:00:36.775 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Registering mariadb_recover_inventory_name from temp file] *****",
                                "Friday 13 March 2026  00:56:57 +0000 (0:00:00.288)       0:00:37.064 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap and master hostnames into facts] ***************",
                                "Friday 13 March 2026  00:56:57 +0000 (0:00:00.400)       0:00:37.464 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Set grastate.dat file from MariaDB container in bootstrap host] ***",
                                "Friday 13 March 2026  00:56:58 +0000 (0:00:00.288)       0:00:37.753 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Starting first MariaDB container] ******************************",
                                "Friday 13 March 2026  00:56:58 +0000 (0:00:00.288)       0:00:38.041 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for first MariaDB container] ******************************",
                                "Friday 13 March 2026  00:56:58 +0000 (0:00:00.284)       0:00:38.326 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Set first MariaDB container as primary] ************************",
                                "Friday 13 March 2026  00:56:59 +0000 (0:00:00.306)       0:00:38.632 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB to become operational] ************************",
                                "Friday 13 March 2026  00:56:59 +0000 (0:00:00.430)       0:00:39.063 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Restart slave MariaDB container(s)] ****************************",
                                "Friday 13 March 2026  00:56:59 +0000 (0:00:00.278)       0:00:39.342 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for slave MariaDB] ****************************************",
                                "Friday 13 March 2026  00:57:02 +0000 (0:00:02.296)       0:00:41.638 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Restart master MariaDB container(s)] ***************************",
                                "Friday 13 March 2026  00:57:02 +0000 (0:00:00.283)       0:00:41.922 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2025.1', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for master mariadb] ***************************************",
                                "Friday 13 March 2026  00:57:04 +0000 (0:00:02.378)       0:00:44.301 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Get container facts] ***************************",
                                "Friday 13 March 2026  00:57:05 +0000 (0:00:00.281)       0:00:44.582 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are missing or not running] ***",
                                "Friday 13 March 2026  00:57:05 +0000 (0:00:00.291)       0:00:44.873 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are unhealthy] **************",
                                "Friday 13 March 2026  00:57:05 +0000 (0:00:00.279)       0:00:45.153 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Friday 13 March 2026  00:57:06 +0000 (0:00:00.579)       0:00:45.733 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Friday 13 March 2026  00:57:06 +0000 (0:00:00.284)       0:00:46.018 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Friday 13 March 2026  00:57:07 +0000 (0:00:00.952)       0:00:46.970 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Friday 13 March 2026  00:57:07 +0000 (0:00:00.398)       0:00:47.369 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Friday 13 March 2026  00:57:08 +0000 (0:00:00.300)       0:00:47.670 ********** ",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Friday 13 March 2026  00:57:18 +0000 (0:00:10.775)       0:00:58.445 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Friday 13 March 2026  00:57:19 +0000 (0:00:00.286)       0:00:58.731 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Friday 13 March 2026  00:57:19 +0000 (0:00:00.404)       0:00:59.136 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Friday 13 March 2026  00:57:19 +0000 (0:00:00.325)       0:00:59.462 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Friday 13 March 2026  00:57:20 +0000 (0:00:00.321)       0:00:59.784 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Friday 13 March 2026  00:57:20 +0000 (0:00:00.326)       0:01:00.110 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:57:21 +0000 (0:00:00.496)       0:01:00.606 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [mariadb : Running MariaDB bootstrap container] ***************************",
                                "Friday 13 March 2026  00:57:21 +0000 (0:00:00.366)       0:01:00.972 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap host name into facts] **************************",
                                "Friday 13 March 2026  00:57:31 +0000 (0:00:09.987)       0:01:10.959 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:57:31 +0000 (0:00:00.116)       0:01:11.076 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Starting first MariaDB container] *******************",
                                "Friday 13 March 2026  00:57:32 +0000 (0:00:00.850)       0:01:11.926 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******",
                                "Friday 13 March 2026  00:57:39 +0000 (0:00:07.389)       0:01:19.315 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******",
                                "Friday 13 March 2026  00:57:41 +0000 (0:00:01.726)       0:01:21.042 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***",
                                "Friday 13 March 2026  00:57:43 +0000 (0:00:02.167)       0:01:23.210 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Friday 13 March 2026  00:57:43 +0000 (0:00:00.132)       0:01:23.342 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Friday 13 March 2026  00:57:44 +0000 (0:00:00.363)       0:01:23.706 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Friday 13 March 2026  00:57:44 +0000 (0:00:00.551)       0:01:24.257 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Friday 13 March 2026  00:57:58 +0000 (0:00:14.076)       0:01:38.334 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Friday 13 March 2026  00:58:13 +0000 (0:00:14.503)       0:01:52.838 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Friday 13 March 2026  00:58:15 +0000 (0:00:01.887)       0:01:54.726 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Friday 13 March 2026  00:58:35 +0000 (0:00:19.917)       0:02:14.643 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Friday 13 March 2026  00:58:44 +0000 (0:00:09.573)       0:02:24.217 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Friday 13 March 2026  00:58:46 +0000 (0:00:01.868)       0:02:26.085 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Friday 13 March 2026  00:58:56 +0000 (0:00:09.776)       0:02:35.862 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Friday 13 March 2026  00:59:00 +0000 (0:00:04.597)       0:02:40.460 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Friday 13 March 2026  00:59:03 +0000 (0:00:02.110)       0:02:42.571 ********** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Friday 13 March 2026  00:59:03 +0000 (0:00:00.473)       0:02:43.045 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Friday 13 March 2026  00:59:06 +0000 (0:00:02.468)       0:02:45.513 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Friday 13 March 2026  00:59:08 +0000 (0:00:02.635)       0:02:48.148 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Friday 13 March 2026  00:59:10 +0000 (0:00:02.132)       0:02:50.281 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Get container facts] ***************************",
                                "Friday 13 March 2026  00:59:12 +0000 (0:00:02.011)       0:02:52.292 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are missing or not running] ***",
                                "Friday 13 March 2026  00:59:17 +0000 (0:00:04.405)       0:02:56.698 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check : mariadb | Fail if containers are unhealthy] **************",
                                "Friday 13 March 2026  00:59:19 +0000 (0:00:02.355)       0:02:59.054 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Friday 13 March 2026  00:59:20 +0000 (0:00:00.523)       0:02:59.577 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Friday 13 March 2026  00:59:22 +0000 (0:00:02.659)       0:03:02.237 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=36  \u001b[0m \u001b[0;33mchanged=17  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=39  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  00:59:23 +0000 (0:00:00.400)       0:03:02.637 ********** ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 34.00s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 24.08s",
                                "mariadb : Check MariaDB service port liveness -------------------------- 10.78s",
                                "mariadb : Running MariaDB bootstrap container --------------------------- 9.99s",
                                "mariadb : Restart MariaDB container ------------------------------------- 9.78s",
                                "mariadb : Starting first MariaDB container ------------------------------ 7.39s",
                                "mariadb : Wait for MariaDB service port liveness ------------------------ 4.60s",
                                "service-check : mariadb | Get container facts --------------------------- 4.41s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 4.18s",
                                "mariadb : Copying over config.json files for services ------------------- 3.84s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 3.76s",
                                "mariadb : Ensuring config directories exist ----------------------------- 3.40s",
                                "Check MariaDB service --------------------------------------------------- 2.86s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 2.69s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 2.66s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 2.66s",
                                "mariadb : Creating mysql monitor user ----------------------------------- 2.64s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 2.52s",
                                "service-check-containers : mariadb | Check containers ------------------- 2.51s",
                                "mariadb : Creating shard root mysql user -------------------------------- 2.47s",
                                "\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 91050273-b89e-432c-a0df-09343bbc34a2 is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Create ceph pools] *******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Friday 13 March 2026  00:58:17 +0000 (0:00:00.537)       0:00:00.537 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Friday 13 March 2026  00:58:18 +0000 (0:00:00.532)       0:00:01.070 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Friday 13 March 2026  00:58:18 +0000 (0:00:00.566)       0:00:01.636 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Friday 13 March 2026  00:58:19 +0000 (0:00:00.274)       0:00:01.911 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Friday 13 March 2026  00:58:19 +0000 (0:00:00.687)       0:00:02.599 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Friday 13 March 2026  00:58:19 +0000 (0:00:00.260)       0:00:02.860 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Friday 13 March 2026  00:58:20 +0000 (0:00:00.286)       0:00:03.146 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Friday 13 March 2026  00:58:20 +0000 (0:00:00.307)       0:00:03.454 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Friday 13 March 2026  00:58:20 +0000 (0:00:00.392)       0:00:03.846 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Friday 13 March 2026  00:58:21 +0000 (0:00:00.249)       0:00:04.096 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Friday 13 March 2026  00:58:21 +0000 (0:00:00.583)       0:00:04.680 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Friday 13 March 2026  00:58:22 +0000 (0:00:00.372)       0:00:05.052 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Friday 13 March 2026  00:58:23 +0000 (0:00:01.808)       0:00:06.861 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Friday 13 March 2026  00:58:24 +0000 (0:00:00.462)       0:00:07.323 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Friday 13 March 2026  00:58:25 +0000 (0:00:00.856)       0:00:08.179 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Friday 13 March 2026  00:58:25 +0000 (0:00:00.267)       0:00:08.446 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '8f8762892bf0', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-03-13 00:58:22.722015', 'end': '2026-03-13 00:58:22.745024', 'delta': '0:00:00.023009', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['8f8762892bf0'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': 'cb0a6fe177a7', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-03-13 00:58:23.348799', 'end': '2026-03-13 00:58:23.374570', 'delta': '0:00:00.025771', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['cb0a6fe177a7'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '73131559e39b', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-03-13 00:58:23.809996', 'end': '2026-03-13 00:58:23.845030', 'delta': '0:00:00.035034', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['73131559e39b'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Friday 13 March 2026  00:58:25 +0000 (0:00:00.177)       0:00:08.623 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Friday 13 March 2026  00:58:26 +0000 (0:00:00.377)       0:00:09.001 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Friday 13 March 2026  00:58:27 +0000 (0:00:01.888)       0:00:10.889 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Friday 13 March 2026  00:58:28 +0000 (0:00:00.280)       0:00:11.170 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Friday 13 March 2026  00:58:28 +0000 (0:00:00.350)       0:00:11.521 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Friday 13 March 2026  00:58:29 +0000 (0:00:00.382)       0:00:11.903 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Friday 13 March 2026  00:58:29 +0000 (0:00:00.108)       0:00:12.012 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Friday 13 March 2026  00:58:29 +0000 (0:00:00.230)       0:00:12.243 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Friday 13 March 2026  00:58:29 +0000 (0:00:00.253)       0:00:12.497 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Friday 13 March 2026  00:58:29 +0000 (0:00:00.279)       0:00:12.776 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Friday 13 March 2026  00:58:30 +0000 (0:00:00.383)       0:00:13.160 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Friday 13 March 2026  00:58:30 +0000 (0:00:00.296)       0:00:13.456 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Friday 13 March 2026  00:58:30 +0000 (0:00:00.262)       0:00:13.719 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Friday 13 March 2026  00:58:31 +0000 (0:00:00.271)       0:00:13.991 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Friday 13 March 2026  00:58:31 +0000 (0:00:00.442)       0:00:14.434 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Friday 13 March 2026  00:58:32 +0000 (0:00:00.555)       0:00:14.989 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62', 'dm-uuid-LVM-gmWIN98WuPHOess4mYE4aaT5b63hd1x6Fdjcij4rFAl0USOwhQSdfqsWQFDnmbd7'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158', 'dm-uuid-LVM-OtMYsV6JsFfkvLjEoQuSzLA6b0b7m713Khduqou86G5KZBKKX31UacsmRHQZufM4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419', 'dm-uuid-LVM-Nm8wQyY0AbE2QjqjXaGDetAadlG8ajrmcM7T3bF3gOHkdeHSTcGQBgSH4nEkpuqW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac', 'dm-uuid-LVM-FKqixNkpGFEPrRS0nQ5XnPYx0vX5HSFcdRpmpLdfuoiTAqmuHK8bnsTflOWLsb2N'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16', 'scsi-SQEMU_QEMU_HARDDISK_f673daef-4b26-4c19-9ad2-6e63015dc391-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--34f0c902--d036--5088--bfac--53928dad7c62-osd--block--34f0c902--d036--5088--bfac--53928dad7c62'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-eakt0p-Bmpf-e6ba-hMhE-jEGF-dzpU-ZS92gF', 'scsi-0QEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1', 'scsi-SQEMU_QEMU_HARDDISK_f4e98a13-908f-47c9-b3df-604b890733f1'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--74a7d81d--0a8f--5a99--882a--be09baf5f158-osd--block--74a7d81d--0a8f--5a99--882a--be09baf5f158'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-wfOa15-1CNh-pWRp-OAvV-Gzjq-pwf5-ixTSF0', 'scsi-0QEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604', 'scsi-SQEMU_QEMU_HARDDISK_77efb09a-3c5f-4577-ba7d-fc82392d2604'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9', 'scsi-SQEMU_QEMU_HARDDISK_45a6e807-92ae-410a-8871-af67928fbfa9'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-27-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16', 'scsi-SQEMU_QEMU_HARDDISK_b0e3959e-31e2-4ee1-98d8-465e3bcb125b-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--14d46521--cd63--5c40--b949--080247a26419-osd--block--14d46521--cd63--5c40--b949--080247a26419'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-hcS0b8-4Opx-63wJ-ftp2-Xaf1-BJNw-mByAGw', 'scsi-0QEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3', 'scsi-SQEMU_QEMU_HARDDISK_7a80a5ff-f3bf-405b-902d-2c7a644821a3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f', 'dm-uuid-LVM-a8eXbdIsz4EtwVcbUWfu1w7Au2MtJoMQgfwhOQXBT2n4rLueplGAIYlwsZwX8BZt'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--27536765--3469--5fa1--a30d--2781c44211ac-osd--block--27536765--3469--5fa1--a30d--2781c44211ac'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Cr9fNM-ADkD-K4Gu-Gs3S-DMs3-biTj-t5Aq1E', 'scsi-0QEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b', 'scsi-SQEMU_QEMU_HARDDISK_f4a99b06-bc15-4835-b50d-6358b82edd6b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4', 'dm-uuid-LVM-YIjL6wTadmMRGkWAGDMPCWNlFVAHlhTY2Puwh6VYhsfrSVwgG0w0Mg9U62ScLyz8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917', 'scsi-SQEMU_QEMU_HARDDISK_de104e6f-6514-4ab5-a08d-43193a3bf917'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-08-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16', 'scsi-SQEMU_QEMU_HARDDISK_bb461c46-8762-4442-a15b-8a2831c93e44-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--a323501f--4860--533e--be9d--aa017f55891f-osd--block--a323501f--4860--533e--be9d--aa017f55891f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-WnjpjE-GGHc-d7K5-ck6F-3Kep-DpqA-6yMjkK', 'scsi-0QEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81', 'scsi-SQEMU_QEMU_HARDDISK_9de382f4-b65a-458b-8cc5-b0eb7440fe81'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--02d8354f--c2be--5171--944a--5a76777090d4-osd--block--02d8354f--c2be--5171--944a--5a76777090d4'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aEqNS3-R2a5-CJiB-bOOS-c69s-NrZh-WBT6wv', 'scsi-0QEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc', 'scsi-SQEMU_QEMU_HARDDISK_9392309c-12b1-408e-a498-5a5dc61ed8cc'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4', 'scsi-SQEMU_QEMU_HARDDISK_da000f05-04bb-42ea-839f-c75cda06f5f4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-03-13-00-03-25-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Friday 13 March 2026  00:58:32 +0000 (0:00:00.539)       0:00:15.529 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Friday 13 March 2026  00:58:33 +0000 (0:00:00.615)       0:00:16.145 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Friday 13 March 2026  00:58:33 +0000 (0:00:00.352)       0:00:16.497 ********** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Friday 13 March 2026  00:58:34 +0000 (0:00:00.553)       0:00:17.051 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Friday 13 March 2026  00:58:34 +0000 (0:00:00.213)       0:00:17.264 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Friday 13 March 2026  00:58:34 +0000 (0:00:00.325)       0:00:17.590 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Friday 13 March 2026  00:58:35 +0000 (0:00:00.353)       0:00:17.944 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Friday 13 March 2026  00:58:35 +0000 (0:00:00.805)       0:00:18.749 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Friday 13 March 2026  00:58:36 +0000 (0:00:00.301)       0:00:19.051 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Friday 13 March 2026  00:58:36 +0000 (0:00:00.555)       0:00:19.606 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Friday 13 March 2026  00:58:36 +0000 (0:00:00.264)       0:00:19.871 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Friday 13 March 2026  00:58:37 +0000 (0:00:00.274)       0:00:20.145 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Friday 13 March 2026  00:58:37 +0000 (0:00:00.271)       0:00:20.417 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Friday 13 March 2026  00:58:38 +0000 (0:00:00.480)       0:00:20.897 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Friday 13 March 2026  00:58:38 +0000 (0:00:00.361)       0:00:21.259 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Friday 13 March 2026  00:58:38 +0000 (0:00:00.343)       0:00:21.602 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Friday 13 March 2026  00:58:39 +0000 (0:00:00.330)       0:00:21.933 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Friday 13 March 2026  00:58:39 +0000 (0:00:00.264)       0:00:22.197 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Friday 13 March 2026  00:58:39 +0000 (0:00:00.396)       0:00:22.594 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Friday 13 March 2026  00:58:40 +0000 (0:00:00.741)       0:00:23.335 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Include tasks from the ceph-osd role] ************************************",
                                "Friday 13 March 2026  00:58:42 +0000 (0:00:01.773)       0:00:25.108 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [create openstack pool(s)] ************************************************",
                                "Friday 13 March 2026  00:58:42 +0000 (0:00:00.323)       0:00:25.432 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [generate keys] ***********************************************************",
                                "Friday 13 March 2026  00:59:28 +0000 (0:00:46.397)       0:01:11.830 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [get keys from monitors] **************************************************",
                                "Friday 13 March 2026  00:59:52 +0000 (0:00:23.640)       0:01:35.470 ********** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [copy ceph key(s) if needed] **********************************************",
                                "Friday 13 March 2026  01:00:04 +0000 (0:00:11.858)       0:01:47.329 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:00:20 +0000 (0:00:16.181)       0:02:03.510 ********** ",
                                "=============================================================================== ",
                                "create openstack pool(s) ----------------------------------------------- 46.40s",
                                "generate keys ---------------------------------------------------------- 23.64s",
                                "copy ceph key(s) if needed --------------------------------------------- 16.18s",
                                "get keys from monitors ------------------------------------------------- 11.86s",
                                "ceph-facts : Get current fsid if cluster is already running ------------- 1.89s",
                                "ceph-facts : Find a running mon container ------------------------------- 1.81s",
                                "ceph-facts : Set_fact ceph_admin_command -------------------------------- 1.77s",
                                "ceph-facts : Check if the ceph mon socket is in-use --------------------- 0.86s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.81s",
                                "ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.74s",
                                "ceph-facts : Check if podman binary is present -------------------------- 0.69s",
                                "ceph-facts : Check if the ceph conf exists ------------------------------ 0.62s",
                                "ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.58s",
                                "ceph-facts : Check if it is atomic host --------------------------------- 0.57s",
                                "ceph-facts : Collect existed devices ------------------------------------ 0.56s",
                                "ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.56s",
                                "ceph-facts : Read osd pool default crush rule --------------------------- 0.55s",
                                "ceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.54s",
                                "ceph-facts : Include facts.yml ------------------------------------------ 0.53s",
                                "ceph-facts : Set_fact _radosgw_address to radosgw_address --------------- 0.48s",
                                "\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c50c0e3e-7019-461b-a0ab-7e7942f81c26 is in state SUCCESS\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:59:27 +0000 (0:00:00.192)       0:00:00.192 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:59:27 +0000 (0:00:00.211)       0:00:00.403 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m",
                                "",
                                "PLAY [Apply role horizon] ******************************************************",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:59:28 +0000 (0:00:00.368)       0:00:00.772 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Ensuring config directories exist] *****************************",
                                "Friday 13 March 2026  00:59:28 +0000 (0:00:00.402)       0:00:01.174 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : Set empty custom policy] ***************************************",
                                "Friday 13 March 2026  00:59:29 +0000 (0:00:01.132)       0:00:02.307 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.344)       0:00:02.651 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.598)       0:00:03.249 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.260)       0:00:03.509 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.094)       0:00:03.603 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:31 +0000 (0:00:00.340)       0:00:03.944 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:31 +0000 (0:00:00.261)       0:00:04.205 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:31 +0000 (0:00:00.117)       0:00:04.322 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:31 +0000 (0:00:00.242)       0:00:04.564 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:32 +0000 (0:00:00.273)       0:00:04.838 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:32 +0000 (0:00:00.224)       0:00:05.063 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:32 +0000 (0:00:00.248)       0:00:05.311 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:32 +0000 (0:00:00.257)       0:00:05.569 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:33 +0000 (0:00:00.114)       0:00:05.683 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:33 +0000 (0:00:00.232)       0:00:05.916 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:33 +0000 (0:00:00.448)       0:00:06.365 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:33 +0000 (0:00:00.122)       0:00:06.488 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:34 +0000 (0:00:00.299)       0:00:06.788 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:34 +0000 (0:00:00.305)       0:00:07.093 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:34 +0000 (0:00:00.171)       0:00:07.265 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:34 +0000 (0:00:00.262)       0:00:07.527 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:35 +0000 (0:00:00.536)       0:00:08.064 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:35 +0000 (0:00:00.175)       0:00:08.239 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:35 +0000 (0:00:00.339)       0:00:08.578 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:36 +0000 (0:00:00.306)       0:00:08.884 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:36 +0000 (0:00:00.128)       0:00:09.013 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:36 +0000 (0:00:00.458)       0:00:09.471 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:37 +0000 (0:00:00.340)       0:00:09.811 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:37 +0000 (0:00:00.114)       0:00:09.926 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Friday 13 March 2026  00:59:37 +0000 (0:00:00.294)       0:00:10.221 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  00:59:37 +0000 (0:00:00.290)       0:00:10.511 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Friday 13 March 2026  00:59:37 +0000 (0:00:00.122)       0:00:10.633 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over config.json files for services] *******************",
                                "Friday 13 March 2026  00:59:38 +0000 (0:00:00.476)       0:00:11.110 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over horizon.conf] *************************************",
                                "Friday 13 March 2026  00:59:40 +0000 (0:00:01.743)       0:00:12.854 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over kolla-settings.py] ********************************",
                                "Friday 13 March 2026  00:59:41 +0000 (0:00:01.638)       0:00:14.493 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom-settings.py] *******************************",
                                "Friday 13 March 2026  00:59:44 +0000 (0:00:02.339)       0:00:16.832 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over existing policy file] *****************************",
                                "Friday 13 March 2026  00:59:46 +0000 (0:00:01.960)       0:00:18.792 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom themes] ************************************",
                                "Friday 13 March 2026  00:59:46 +0000 (0:00:00.291)       0:00:19.084 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:59:46 +0000 (0:00:00.299)       0:00:19.384 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  00:59:47 +0000 (0:00:00.757)       0:00:20.141 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:59:49 +0000 (0:00:01.554)       0:00:21.695 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  00:59:49 +0000 (0:00:00.622)       0:00:22.317 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : horizon | Check containers] *******************",
                                "Friday 13 March 2026  00:59:50 +0000 (0:00:00.958)       0:00:23.275 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : horizon | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  00:59:52 +0000 (0:00:01.425)       0:00:24.701 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  00:59:52 +0000 (0:00:00.302)       0:00:25.004 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2025.1', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_VENUS': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin', 'option httpchk'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:59:53 +0000 (0:00:00.865)       0:00:25.869 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Friday 13 March 2026  00:59:53 +0000 (0:00:00.407)       0:00:26.276 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database] *************************************",
                                "Friday 13 March 2026  00:59:54 +0000 (0:00:00.465)       0:00:26.741 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database user and setting permissions] ********",
                                "Friday 13 March 2026  00:59:56 +0000 (0:00:02.800)       0:00:29.542 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Running Horizon bootstrap container] ***************************",
                                "Friday 13 March 2026  00:59:59 +0000 (0:00:02.595)       0:00:32.138 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:14.887)       0:00:47.026 ********** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:00.077)       0:00:47.104 ********** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:00.244)       0:00:47.349 ********** ",
                                "",
                                "RUNNING HANDLER [horizon : Restart horizon container] **************************",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:00.064)       0:00:47.413 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=26  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=16  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=16  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:00:55 +0000 (0:00:41.041)       0:01:28.455 ********** ",
                                "=============================================================================== ",
                                "horizon : Restart horizon container ------------------------------------ 41.04s",
                                "horizon : Running Horizon bootstrap container -------------------------- 14.89s",
                                "horizon : Creating Horizon database ------------------------------------- 2.80s",
                                "horizon : Creating Horizon database user and setting permissions -------- 2.60s",
                                "horizon : Copying over kolla-settings.py -------------------------------- 2.34s",
                                "horizon : Copying over custom-settings.py ------------------------------- 1.96s",
                                "horizon : Copying over config.json files for services ------------------- 1.74s",
                                "horizon : Copying over horizon.conf ------------------------------------- 1.64s",
                                "service-cert-copy : horizon | Copying over extra CA certificates -------- 1.55s",
                                "service-check-containers : horizon | Check containers ------------------- 1.43s",
                                "horizon : Ensuring config directories exist ----------------------------- 1.13s",
                                "service-cert-copy : horizon | Copying over backend internal TLS key ----- 0.96s",
                                "service-check-containers : Include tasks -------------------------------- 0.87s",
                                "horizon : include_tasks ------------------------------------------------- 0.76s",
                                "service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.62s",
                                "horizon : include_tasks ------------------------------------------------- 0.60s",
                                "horizon : Update policy file name --------------------------------------- 0.54s",
                                "horizon : Update custom policy file name -------------------------------- 0.48s",
                                "horizon : include_tasks ------------------------------------------------- 0.47s",
                                "horizon : Update custom policy file name -------------------------------- 0.46s",
                                "\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2a714bc0-332b-4ae9-ad61-c4404f9ecf52 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cea88b5-7485-4a64-8947-fe6ffa2e7647 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy ceph keys to the configuration repository] **************************",
                                "",
                                "TASK [Check if ceph keys exist] ************************************************",
                                "Friday 13 March 2026  01:00:25 +0000 (0:00:00.141)       0:00:00.141 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Fetch all ceph keys] *****************************************************",
                                "Friday 13 March 2026  01:00:29 +0000 (0:00:04.100)       0:00:04.241 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Create share directory] **************************************************",
                                "Friday 13 March 2026  01:00:32 +0000 (0:00:03.523)       0:00:07.765 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the share directory] **********************************",
                                "Friday 13 March 2026  01:00:33 +0000 (0:00:00.919)       0:00:08.684 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Check if target directories exist] ***************************************",
                                "Friday 13 March 2026  01:00:46 +0000 (0:00:12.684)       0:00:21.369 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the configuration directory] **************************",
                                "Friday 13 March 2026  01:00:49 +0000 (0:00:02.700)       0:00:24.069 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:00:54 +0000 (0:00:05.838)       0:00:29.908 ********** ",
                                "=============================================================================== ",
                                "Write ceph keys to the share directory --------------------------------- 12.68s",
                                "Write ceph keys to the configuration directory -------------------------- 5.84s",
                                "Check if ceph keys exist ------------------------------------------------ 4.10s",
                                "Fetch all ceph keys ----------------------------------------------------- 3.52s",
                                "Check if target directories exist --------------------------------------- 2.70s",
                                "Create share directory -------------------------------------------------- 0.92s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  00:59:27 +0000 (0:00:00.214)       0:00:00.214 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  00:59:27 +0000 (0:00:00.211)       0:00:00.426 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Apply role keystone] *****************************************************",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:59:28 +0000 (0:00:00.360)       0:00:00.786 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Ensuring config directories exist] ****************************",
                                "Friday 13 March 2026  00:59:28 +0000 (0:00:00.431)       0:00:01.217 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Check if policies shall be overwritten] ***********************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:01.940)       0:00:03.158 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Set keystone policy file] *************************************",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.126)       0:00:03.284 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Check if Keystone domain-specific config is supplied] *********",
                                "Friday 13 March 2026  00:59:30 +0000 (0:00:00.342)       0:00:03.626 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:59:31 +0000 (0:00:00.692)       0:00:04.319 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over extra CA certificates] *******",
                                "Friday 13 March 2026  00:59:32 +0000 (0:00:00.401)       0:00:04.721 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  00:59:35 +0000 (0:00:03.294)       0:00:08.015 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****",
                                "Friday 13 March 2026  00:59:36 +0000 (0:00:00.619)       0:00:08.635 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over config.json files for services] ******************",
                                "Friday 13 March 2026  00:59:36 +0000 (0:00:00.757)       0:00:09.393 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone.conf] ***********************************",
                                "Friday 13 March 2026  00:59:40 +0000 (0:00:03.852)       0:00:13.246 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying keystone-startup script for keystone] *****************",
                                "Friday 13 March 2026  00:59:45 +0000 (0:00:04.952)       0:00:18.198 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Create Keystone domain-specific config directory] *************",
                                "Friday 13 March 2026  00:59:47 +0000 (0:00:01.563)       0:00:19.761 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Get file list in custom domains folder] ***********************",
                                "Friday 13 March 2026  00:59:47 +0000 (0:00:00.539)       0:00:20.301 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying Keystone Domain specific settings] ********************",
                                "Friday 13 March 2026  00:59:47 +0000 (0:00:00.325)       0:00:20.627 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over existing policy file] ****************************",
                                "Friday 13 March 2026  00:59:48 +0000 (0:00:00.486)       0:00:21.113 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  00:59:49 +0000 (0:00:00.580)       0:00:21.694 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over wsgi-keystone.conf] ******************************",
                                "Friday 13 March 2026  00:59:49 +0000 (0:00:00.301)       0:00:21.995 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "",
                                "TASK [keystone : Checking whether keystone-paste.ini file exists] **************",
                                "Friday 13 March 2026  00:59:51 +0000 (0:00:01.671)       0:00:23.667 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone-paste.ini] ******************************",
                                "Friday 13 March 2026  00:59:51 +0000 (0:00:00.872)       0:00:24.539 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Generate the required cron jobs for the node] *****************",
                                "Friday 13 March 2026  00:59:52 +0000 (0:00:00.633)       0:00:25.172 ********** ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***",
                                "Friday 13 March 2026  00:59:53 +0000 (0:00:01.073)       0:00:26.246 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-fernet] ****************************",
                                "Friday 13 March 2026  00:59:53 +0000 (0:00:00.273)       0:00:26.520 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-ssh] *******************************",
                                "Friday 13 March 2026  01:00:02 +0000 (0:00:08.695)       0:00:35.216 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "",
                                "TASK [service-check-containers : keystone | Check containers] ******************",
                                "Friday 13 March 2026  01:00:05 +0000 (0:00:02.664)       0:00:37.880 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : keystone | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:00:07 +0000 (0:00:02.422)       0:00:40.303 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:00:08 +0000 (0:00:00.360)       0:00:40.663 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2025.1', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin', 'option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2025.1', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2025.1', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:00:08 +0000 (0:00:00.917)       0:00:41.581 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Creating keystone database] ***********************************",
                                "Friday 13 March 2026  01:00:09 +0000 (0:00:00.288)       0:00:41.869 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating Keystone database user and setting permissions] ******",
                                "Friday 13 March 2026  01:00:11 +0000 (0:00:02.076)       0:00:43.946 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Checking for any running keystone_fernet containers] **********",
                                "Friday 13 March 2026  01:00:13 +0000 (0:00:02.246)       0:00:46.192 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Group nodes where keystone_fernet is running] *****************",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:00.861)       0:00:47.054 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***",
                                "Friday 13 March 2026  01:00:14 +0000 (0:00:00.323)       0:00:47.378 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone bootstrap container] *************************",
                                "Friday 13 March 2026  01:00:15 +0000 (0:00:00.681)       0:00:48.059 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone fernet bootstrap container] ******************",
                                "Friday 13 March 2026  01:00:29 +0000 (0:00:13.723)       0:01:01.783 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:00:38 +0000 (0:00:09.136)       0:01:10.919 ********** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:10.970 ********** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:11.020 ********** ",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-ssh container] ********************",
                                "Friday 13 March 2026  01:00:38 +0000 (0:00:00.050)       0:01:11.071 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-fernet container] *****************",
                                "Friday 13 March 2026  01:00:51 +0000 (0:00:13.421)       0:01:24.492 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone container] ************************",
                                "Friday 13 March 2026  01:01:01 +0000 (0:00:09.295)       0:01:33.788 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:01:07 +0000 (0:00:06.275)       0:01:40.063 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Waiting for Keystone SSH port to be UP] ***********************",
                                "Friday 13 March 2026  01:01:07 +0000 (0:00:00.505)       0:01:40.568 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Run key distribution] *****************************************",
                                "Friday 13 March 2026  01:01:08 +0000 (0:00:00.910)       0:01:41.478 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating admin project, user, role, service, and endpoint] ****",
                                "Friday 13 March 2026  01:01:10 +0000 (0:00:01.481)       0:01:42.960 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating/deleting services] *************",
                                "Friday 13 March 2026  01:01:21 +0000 (0:00:11.495)       0:01:54.456 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating/deleting endpoints] ************",
                                "Friday 13 March 2026  01:01:25 +0000 (0:00:04.152)       0:01:58.608 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating projects] **********************",
                                "Friday 13 March 2026  01:01:32 +0000 (0:00:07.021)       0:02:05.629 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating users] *************************",
                                "Friday 13 March 2026  01:01:33 +0000 (0:00:00.113)       0:02:05.742 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating roles] *************************",
                                "Friday 13 March 2026  01:01:33 +0000 (0:00:00.111)       0:02:05.853 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Granting/revoking user roles] ***********",
                                "Friday 13 March 2026  01:01:33 +0000 (0:00:00.115)       0:02:05.969 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating default user role] ***********************************",
                                "Friday 13 March 2026  01:01:33 +0000 (0:00:00.302)       0:02:06.271 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:01:36 +0000 (0:00:03.214)       0:02:09.486 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:01:37 +0000 (0:00:00.366)       0:02:09.852 ********** ",
                                "=============================================================================== ",
                                "keystone : Running Keystone bootstrap container ------------------------ 13.72s",
                                "keystone : Restart keystone-ssh container ------------------------------ 13.42s",
                                "keystone : Creating admin project, user, role, service, and endpoint --- 11.50s",
                                "keystone : Restart keystone-fernet container ---------------------------- 9.30s",
                                "keystone : Running Keystone fernet bootstrap container ------------------ 9.14s",
                                "keystone : Copying files for keystone-fernet ---------------------------- 8.70s",
                                "service-ks-register : keystone | Creating/deleting endpoints ------------ 7.02s",
                                "keystone : Restart keystone container ----------------------------------- 6.28s",
                                "keystone : Copying over keystone.conf ----------------------------------- 4.95s",
                                "service-ks-register : keystone | Creating/deleting services ------------- 4.15s",
                                "keystone : Copying over config.json files for services ------------------ 3.85s",
                                "service-cert-copy : keystone | Copying over extra CA certificates ------- 3.29s",
                                "keystone : Creating default user role ----------------------------------- 3.21s",
                                "keystone : Copying files for keystone-ssh ------------------------------- 2.66s",
                                "service-check-containers : keystone | Check containers ------------------ 2.42s",
                                "keystone : Creating Keystone database user and setting permissions ------ 2.25s",
                                "keystone : Creating keystone database ----------------------------------- 2.08s",
                                "keystone : Ensuring config directories exist ---------------------------- 1.94s",
                                "keystone : Copying over wsgi-keystone.conf ------------------------------ 1.67s",
                                "keystone : Copying keystone-startup script for keystone ----------------- 1.56s",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bca6831c-77be-4e16-a7a4-eb7b5aba56c7 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7500dcc-4e8e-4052-a998-09762d1ec6f5 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Friday 13 March 2026  01:00:59 +0000 (0:00:00.217)       0:00:00.217 ********** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Friday 13 March 2026  01:00:59 +0000 (0:00:00.209)       0:00:00.427 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Friday 13 March 2026  01:01:00 +0000 (0:00:01.090)       0:00:01.517 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Friday 13 March 2026  01:01:01 +0000 (0:00:01.160)       0:00:02.677 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Friday 13 March 2026  01:01:02 +0000 (0:00:00.708)       0:00:03.385 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Friday 13 March 2026  01:01:03 +0000 (0:00:00.769)       0:00:04.155 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Friday 13 March 2026  01:01:39 +0000 (0:00:36.490)       0:00:40.645 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Friday 13 March 2026  01:01:43 +0000 (0:00:04.228)       0:00:44.873 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Friday 13 March 2026  01:01:44 +0000 (0:00:00.358)       0:00:45.232 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Friday 13 March 2026  01:01:44 +0000 (0:00:00.095)       0:00:45.328 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******",
                                "Friday 13 March 2026  01:01:44 +0000 (0:00:00.364)       0:00:45.693 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***",
                                "Friday 13 March 2026  01:01:45 +0000 (0:00:01.127)       0:00:46.821 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******",
                                "Friday 13 March 2026  01:01:46 +0000 (0:00:00.548)       0:00:47.369 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****",
                                "Friday 13 March 2026  01:01:46 +0000 (0:00:00.544)       0:00:47.914 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:01:47 +0000 (0:00:01.177)       0:00:49.092 ********** ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Manage cephclient service ------------------ 36.49s",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 4.23s",
                                "osism.services.cephclient : Copy bash completion scripts ---------------- 1.18s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.16s",
                                "osism.services.cephclient : Restart cephclient service ------------------ 1.13s",
                                "osism.services.cephclient : Create required directories ----------------- 1.09s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.77s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.71s",
                                "osism.services.cephclient : Ensure that all containers are up ----------- 0.55s",
                                "osism.services.cephclient : Wait for an healthy service ----------------- 0.54s",
                                "osism.services.cephclient : Include rook task --------------------------- 0.36s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.36s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.21s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.10s",
                                "",
                                "",
                                "PLAY [Download ironic ipa images] **********************************************",
                                "",
                                "TASK [Ensure the destination directory exists] *********************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.115)       0:00:00.115 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent initramfs] *****************************************",
                                "Friday 13 March 2026  01:01:42 +0000 (0:00:01.068)       0:00:01.183 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent kernel] ********************************************",
                                "Friday 13 March 2026  01:02:12 +0000 (0:00:29.782)       0:00:30.966 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (3 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent kernel (2 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:03:03 +0000 (0:00:50.676)       0:01:21.642 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:03:03 +0000 (0:00:00.281)       0:01:21.924 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ironic_False)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True\u001b[0m",
                                "",
                                "PLAY [Apply role ironic] *******************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:03:04 +0000 (0:00:01.068)       0:01:22.992 ********** ",
                                "=============================================================================== ",
                                "Download ironic-agent kernel ------------------------------------------- 50.68s",
                                "Download ironic-agent initramfs ---------------------------------------- 29.78s",
                                "Group hosts based on enabled services ----------------------------------- 1.07s",
                                "Ensure the destination directory exists --------------------------------- 1.07s",
                                "Group hosts based on Kolla action --------------------------------------- 0.28s",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 531c82e3-d5d0-421a-a8b7-77e779e7e3a7 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 93b4d2ad-514f-4039-905a-e4536ffe3f59 is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Bootstraph ceph dashboard] ***********************************************",
                                "",
                                "TASK [Disable the ceph dashboard] **********************************************",
                                "Friday 13 March 2026  01:01:52 +0000 (0:00:00.264)       0:00:00.264 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/ssl to false] ******************************************",
                                "Friday 13 March 2026  01:01:53 +0000 (0:00:01.547)       0:00:01.811 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_port to 7000] ***********************************",
                                "Friday 13 March 2026  01:01:54 +0000 (0:00:00.952)       0:00:02.764 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************",
                                "Friday 13 March 2026  01:01:55 +0000 (0:00:00.953)       0:00:03.718 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_behaviour to error] ****************************",
                                "Friday 13 March 2026  01:01:56 +0000 (0:00:01.025)       0:00:04.743 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_error_status_code to 404] **********************",
                                "Friday 13 March 2026  01:01:57 +0000 (0:00:00.943)       0:00:05.687 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable the ceph dashboard] ***********************************************",
                                "Friday 13 March 2026  01:01:58 +0000 (0:00:00.970)       0:00:06.657 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Write ceph_dashboard_password to temporary file] *************************",
                                "Friday 13 March 2026  01:02:00 +0000 (0:00:01.196)       0:00:07.854 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create admin user] *******************************************************",
                                "Friday 13 March 2026  01:02:01 +0000 (0:00:01.071)       0:00:08.925 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Remove temporary file for ceph_dashboard_password] ***********************",
                                "Friday 13 March 2026  01:02:56 +0000 (0:00:55.493)       0:01:04.418 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Friday 13 March 2026  01:02:56 +0000 (0:00:00.168)       0:01:04.587 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Friday 13 March 2026  01:02:58 +0000 (0:00:01.328)       0:01:05.916 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Friday 13 March 2026  01:03:09 +0000 (0:00:11.596)       0:01:17.512 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:03:20 +0000 (0:00:11.218)       0:01:28.731 ********** ",
                                "=============================================================================== ",
                                "Create admin user ------------------------------------------------------ 55.49s",
                                "Restart ceph manager service ------------------------------------------- 24.14s",
                                "Disable the ceph dashboard ---------------------------------------------- 1.55s",
                                "Enable the ceph dashboard ----------------------------------------------- 1.20s",
                                "Write ceph_dashboard_password to temporary file ------------------------- 1.07s",
                                "Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.03s",
                                "Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 0.97s",
                                "Set mgr/dashboard/server_port to 7000 ----------------------------------- 0.95s",
                                "Set mgr/dashboard/ssl to false ------------------------------------------ 0.95s",
                                "Set mgr/dashboard/standby_behaviour to error ---------------------------- 0.94s",
                                "Remove temporary file for ceph_dashboard_password ----------------------- 0.17s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.198)       0:00:00.198 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.220)       0:00:00.418 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m",
                                "",
                                "PLAY [Apply role barbican] *****************************************************",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:01:42 +0000 (0:00:00.644)       0:00:01.062 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating/deleting services] *************",
                                "Friday 13 March 2026  01:01:43 +0000 (0:00:00.513)       0:00:01.576 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating/deleting endpoints] ************",
                                "Friday 13 March 2026  01:01:47 +0000 (0:00:04.662)       0:00:06.239 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating projects] **********************",
                                "Friday 13 March 2026  01:01:54 +0000 (0:00:06.502)       0:00:12.741 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating users] *************************",
                                "Friday 13 March 2026  01:01:58 +0000 (0:00:03.863)       0:00:16.605 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating roles] *************************",
                                "Friday 13 March 2026  01:02:02 +0000 (0:00:04.254)       0:00:20.859 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Granting/revoking user roles] ***********",
                                "Friday 13 March 2026  01:02:17 +0000 (0:00:14.833)       0:00:35.692 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring config directories exist] ****************************",
                                "Friday 13 March 2026  01:02:21 +0000 (0:00:04.336)       0:00:40.029 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring vassals config directories exist] ********************",
                                "Friday 13 March 2026  01:02:23 +0000 (0:00:02.269)       0:00:42.298 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m",
                                "",
                                "TASK [barbican : Check if policies shall be overwritten] ***********************",
                                "Friday 13 March 2026  01:02:25 +0000 (0:00:01.257)       0:00:43.555 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Set barbican policy file] *************************************",
                                "Friday 13 March 2026  01:02:25 +0000 (0:00:00.108)       0:00:43.664 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:02:25 +0000 (0:00:00.396)       0:00:44.061 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over extra CA certificates] *******",
                                "Friday 13 March 2026  01:02:26 +0000 (0:00:00.460)       0:00:44.521 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:02:29 +0000 (0:00:03.130)       0:00:47.651 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****",
                                "Friday 13 March 2026  01:02:30 +0000 (0:00:01.220)       0:00:48.872 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over config.json files for services] ******************",
                                "Friday 13 March 2026  01:02:31 +0000 (0:00:01.339)       0:00:50.218 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api.ini] ********************************",
                                "Friday 13 March 2026  01:02:35 +0000 (0:00:03.630)       0:00:53.849 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [barbican : Checking whether barbican-api-paste.ini file exists] **********",
                                "Friday 13 March 2026  01:02:38 +0000 (0:00:02.887)       0:00:56.737 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api-paste.ini] **************************",
                                "Friday 13 March 2026  01:02:39 +0000 (0:00:01.179)       0:00:57.916 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican.conf] ***********************************",
                                "Friday 13 March 2026  01:02:40 +0000 (0:00:00.580)       0:00:58.496 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over existing policy file] ****************************",
                                "Friday 13 March 2026  01:02:49 +0000 (0:00:09.058)       0:01:07.555 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : barbican | Check containers] ******************",
                                "Friday 13 March 2026  01:02:50 +0000 (0:00:01.449)       0:01:09.004 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : barbican | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:02:54 +0000 (0:00:04.118)       0:01:13.123 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:02:55 +0000 (0:00:00.857)       0:01:13.981 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2025.1', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2025.1', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2025.1', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:02:56 +0000 (0:00:00.811)       0:01:14.793 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database] ***********************************",
                                "Friday 13 March 2026  01:02:57 +0000 (0:00:00.995)       0:01:15.789 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database user and setting permissions] ******",
                                "Friday 13 March 2026  01:02:59 +0000 (0:00:02.364)       0:01:18.154 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Running barbican bootstrap container] *************************",
                                "Friday 13 March 2026  01:03:01 +0000 (0:00:02.044)       0:01:20.199 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:03:13 +0000 (0:00:11.456)       0:01:31.655 ********** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:03:13 +0000 (0:00:00.144)       0:01:31.801 ********** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Friday 13 March 2026  01:03:13 +0000 (0:00:00.103)       0:01:31.904 ********** ",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-api container] ********************",
                                "Friday 13 March 2026  01:03:13 +0000 (0:00:00.142)       0:01:32.047 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******",
                                "Friday 13 March 2026  01:03:24 +0000 (0:00:11.036)       0:01:43.083 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-worker container] *****************",
                                "Friday 13 March 2026  01:03:29 +0000 (0:00:04.877)       0:01:47.961 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:03:34 +0000 (0:00:05.069)       0:01:53.030 ********** ",
                                "=============================================================================== ",
                                "service-ks-register : barbican | Creating roles ------------------------ 14.83s",
                                "barbican : Running barbican bootstrap container ------------------------ 11.46s",
                                "barbican : Restart barbican-api container ------------------------------ 11.04s",
                                "barbican : Copying over barbican.conf ----------------------------------- 9.06s",
                                "service-ks-register : barbican | Creating/deleting endpoints ------------ 6.50s",
                                "barbican : Restart barbican-worker container ---------------------------- 5.07s",
                                "barbican : Restart barbican-keystone-listener container ----------------- 4.88s",
                                "service-ks-register : barbican | Creating/deleting services ------------- 4.66s",
                                "service-ks-register : barbican | Granting/revoking user roles ----------- 4.34s",
                                "service-ks-register : barbican | Creating users ------------------------- 4.25s",
                                "service-check-containers : barbican | Check containers ------------------ 4.12s",
                                "service-ks-register : barbican | Creating projects ---------------------- 3.86s",
                                "barbican : Copying over config.json files for services ------------------ 3.63s",
                                "service-cert-copy : barbican | Copying over extra CA certificates ------- 3.13s",
                                "barbican : Copying over barbican-api.ini -------------------------------- 2.89s",
                                "barbican : Creating barbican database ----------------------------------- 2.36s",
                                "barbican : Ensuring config directories exist ---------------------------- 2.27s",
                                "barbican : Creating barbican database user and setting permissions ------ 2.05s",
                                "barbican : Copying over existing policy file ---------------------------- 1.45s",
                                "service-cert-copy : barbican | Copying over backend internal TLS key ---- 1.35s",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:03:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 513bb898-6998-4cbf-8e48-5c2b29119b2e is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:00.542)       0:00:00.542 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:00.349)       0:00:00.892 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m",
                                "",
                                "PLAY [Apply role placement] ****************************************************",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:00.373)       0:00:01.265 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating/deleting services] ************",
                                "Friday 13 March 2026  01:03:12 +0000 (0:00:00.486)       0:00:01.752 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating/deleting endpoints] ***********",
                                "Friday 13 March 2026  01:03:15 +0000 (0:00:03.256)       0:00:05.008 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating projects] *********************",
                                "Friday 13 March 2026  01:03:21 +0000 (0:00:06.138)       0:00:11.147 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating users] ************************",
                                "Friday 13 March 2026  01:03:24 +0000 (0:00:03.327)       0:00:14.475 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating roles] ************************",
                                "Friday 13 March 2026  01:03:28 +0000 (0:00:03.632)       0:00:18.107 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Granting/revoking user roles] **********",
                                "Friday 13 March 2026  01:03:31 +0000 (0:00:03.332)       0:00:21.439 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:03:36 +0000 (0:00:04.163)       0:00:25.603 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Ensuring config directories exist] ***************************",
                                "Friday 13 March 2026  01:03:36 +0000 (0:00:00.540)       0:00:26.143 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [placement : Check if policies shall be overwritten] **********************",
                                "Friday 13 March 2026  01:03:37 +0000 (0:00:00.937)       0:00:27.081 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Set placement policy file] ***********************************",
                                "Friday 13 March 2026  01:03:37 +0000 (0:00:00.131)       0:00:27.213 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:03:38 +0000 (0:00:00.753)       0:00:27.966 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over extra CA certificates] ******",
                                "Friday 13 March 2026  01:03:39 +0000 (0:00:00.825)       0:00:28.791 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:03:40 +0000 (0:00:01.535)       0:00:30.327 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS key] ***",
                                "Friday 13 March 2026  01:03:41 +0000 (0:00:01.007)       0:00:31.335 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over config.json files for services] *****************",
                                "Friday 13 March 2026  01:03:43 +0000 (0:00:01.209)       0:00:32.544 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement.conf] *********************************",
                                "Friday 13 March 2026  01:03:44 +0000 (0:00:01.781)       0:00:34.325 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement-api wsgi configuration] ***************",
                                "Friday 13 March 2026  01:03:48 +0000 (0:00:03.272)       0:00:37.597 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Configure uWSGI for Placement] *******************************************",
                                "Friday 13 March 2026  01:03:48 +0000 (0:00:00.409)       0:00:38.007 ********** ",
                                "\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-uwsgi-config : Copying over placement-api uWSGI config] **********",
                                "Friday 13 March 2026  01:03:49 +0000 (0:00:01.130)       0:00:39.137 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [placement : Copying over migrate-db.rc.j2 configuration] *****************",
                                "Friday 13 March 2026  01:03:51 +0000 (0:00:02.123)       0:00:41.260 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over existing policy file] ***************************",
                                "Friday 13 March 2026  01:03:53 +0000 (0:00:01.320)       0:00:42.581 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : placement | Check containers] *****************",
                                "Friday 13 March 2026  01:03:53 +0000 (0:00:00.580)       0:00:43.162 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : placement | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:03:54 +0000 (0:00:01.192)       0:00:44.354 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:03:55 +0000 (0:00:00.460)       0:00:44.815 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'wsgi': 'placement.wsgi.api:application', 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk GET /']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases] ********************************",
                                "Friday 13 March 2026  01:03:55 +0000 (0:00:00.580)       0:00:45.395 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases user and setting permissions] ***",
                                "Friday 13 March 2026  01:03:58 +0000 (0:00:02.198)       0:00:47.593 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Running placement bootstrap container] ***********************",
                                "Friday 13 March 2026  01:04:00 +0000 (0:00:02.294)       0:00:49.888 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:04:13 +0000 (0:00:12.858)       0:01:02.746 ********** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:04:13 +0000 (0:00:00.060)       0:01:02.807 ********** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:04:13 +0000 (0:00:00.249)       0:01:03.056 ********** ",
                                "",
                                "RUNNING HANDLER [placement : Restart placement-api container] ******************",
                                "Friday 13 March 2026  01:04:13 +0000 (0:00:00.071)       0:01:03.128 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:04:18 +0000 (0:00:04.385)       0:01:07.513 ********** ",
                                "=============================================================================== ",
                                "placement : Running placement bootstrap container ---------------------- 12.86s",
                                "service-ks-register : placement | Creating/deleting endpoints ----------- 6.14s",
                                "placement : Restart placement-api container ----------------------------- 4.39s",
                                "service-ks-register : placement | Granting/revoking user roles ---------- 4.16s",
                                "service-ks-register : placement | Creating users ------------------------ 3.63s",
                                "service-ks-register : placement | Creating roles ------------------------ 3.33s",
                                "service-ks-register : placement | Creating projects --------------------- 3.33s",
                                "placement : Copying over placement.conf --------------------------------- 3.27s",
                                "service-ks-register : placement | Creating/deleting services ------------ 3.26s",
                                "placement : Creating placement databases user and setting permissions --- 2.29s",
                                "placement : Creating placement databases -------------------------------- 2.20s",
                                "service-uwsgi-config : Copying over placement-api uWSGI config ---------- 2.12s",
                                "placement : Copying over config.json files for services ----------------- 1.78s",
                                "service-cert-copy : placement | Copying over extra CA certificates ------ 1.54s",
                                "placement : Copying over migrate-db.rc.j2 configuration ----------------- 1.32s",
                                "service-cert-copy : placement | Copying over backend internal TLS key --- 1.21s",
                                "service-check-containers : placement | Check containers ----------------- 1.19s",
                                "Configure uWSGI for Placement ------------------------------------------- 1.13s",
                                "service-cert-copy : placement | Copying over backend internal TLS certificate --- 1.01s",
                                "placement : Ensuring config directories exist --------------------------- 0.94s",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1d7c6370-3618-4d75-8b91-deb0feaaf3b4 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.207)       0:00:00.207 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.237)       0:00:00.444 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m",
                                "",
                                "PLAY [Apply role designate] ****************************************************",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:01:42 +0000 (0:00:00.349)       0:00:00.794 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating/deleting services] ************",
                                "Friday 13 March 2026  01:01:42 +0000 (0:00:00.652)       0:00:01.447 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating/deleting endpoints] ***********",
                                "Friday 13 March 2026  01:01:47 +0000 (0:00:04.703)       0:00:06.150 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating projects] *********************",
                                "Friday 13 March 2026  01:01:53 +0000 (0:00:06.304)       0:00:12.455 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating users] ************************",
                                "Friday 13 March 2026  01:01:57 +0000 (0:00:03.534)       0:00:15.989 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating roles] ************************",
                                "Friday 13 March 2026  01:02:01 +0000 (0:00:04.423)       0:00:20.413 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Granting/revoking user roles] **********",
                                "Friday 13 March 2026  01:02:05 +0000 (0:00:03.373)       0:00:23.786 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m",
                                "",
                                "TASK [designate : Ensuring config directories exist] ***************************",
                                "Friday 13 March 2026  01:02:09 +0000 (0:00:03.846)       0:00:27.633 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Check if policies shall be overwritten] **********************",
                                "Friday 13 March 2026  01:02:11 +0000 (0:00:02.528)       0:00:30.161 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Set designate policy file] ***********************************",
                                "Friday 13 March 2026  01:02:11 +0000 (0:00:00.125)       0:00:30.287 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:02:12 +0000 (0:00:00.254)       0:00:30.542 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over extra CA certificates] ******",
                                "Friday 13 March 2026  01:02:12 +0000 (0:00:00.618)       0:00:31.160 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:02:18 +0000 (0:00:06.050)       0:00:37.211 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS key] ***",
                                "Friday 13 March 2026  01:02:20 +0000 (0:00:01.510)       0:00:38.721 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Copying over config.json files for services] *****************",
                                "Friday 13 March 2026  01:02:22 +0000 (0:00:02.688)       0:00:41.410 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over designate.conf] *********************************",
                                "Friday 13 March 2026  01:02:29 +0000 (0:00:06.539)       0:00:47.949 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over pools.yaml] *************************************",
                                "Friday 13 March 2026  01:02:47 +0000 (0:00:17.558)       0:01:05.508 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over named.conf] *************************************",
                                "Friday 13 March 2026  01:02:53 +0000 (0:00:06.485)       0:01:11.994 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.conf] **************************************",
                                "Friday 13 March 2026  01:02:56 +0000 (0:00:03.119)       0:01:15.113 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.key] ***************************************",
                                "Friday 13 March 2026  01:03:00 +0000 (0:00:03.817)       0:01:18.931 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:03:03 +0000 (0:00:02.919)       0:01:21.850 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over existing policy file] ***************************",
                                "Friday 13 March 2026  01:03:04 +0000 (0:00:00.951)       0:01:22.802 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : designate | Check containers] *****************",
                                "Friday 13 March 2026  01:03:05 +0000 (0:00:01.239)       0:01:24.041 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : designate | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:06.103)       0:01:30.144 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:00.294)       0:01:30.439 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2025.1', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2025.1', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2025.1', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2025.1', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2025.1', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2025.1', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:03:14 +0000 (0:00:02.174)       0:01:32.613 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases] ********************************",
                                "Friday 13 March 2026  01:03:14 +0000 (0:00:00.236)       0:01:32.850 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases user and setting permissions] ***",
                                "Friday 13 March 2026  01:03:16 +0000 (0:00:01.896)       0:01:34.747 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m",
                                "",
                                "TASK [designate : Running Designate bootstrap container] ***********************",
                                "Friday 13 March 2026  01:03:18 +0000 (0:00:02.129)       0:01:36.877 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:03:32 +0000 (0:00:14.152)       0:01:51.029 ********** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:03:32 +0000 (0:00:00.050)       0:01:51.080 ********** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:03:32 +0000 (0:00:00.093)       0:01:51.174 ********** ",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********",
                                "Friday 13 March 2026  01:03:32 +0000 (0:00:00.108)       0:01:51.283 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-api container] ******************",
                                "Friday 13 March 2026  01:03:41 +0000 (0:00:08.494)       0:01:59.777 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-central container] **************",
                                "Friday 13 March 2026  01:03:52 +0000 (0:00:11.136)       0:02:10.913 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-producer container] *************",
                                "Friday 13 March 2026  01:03:57 +0000 (0:00:05.265)       0:02:16.179 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-mdns container] *****************",
                                "Friday 13 March 2026  01:04:02 +0000 (0:00:04.621)       0:02:20.800 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-worker container] ***************",
                                "Friday 13 March 2026  01:04:10 +0000 (0:00:08.110)       0:02:28.910 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Non-destructive DNS pools update] ****************************",
                                "Friday 13 March 2026  01:04:15 +0000 (0:00:05.175)       0:02:34.086 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=25  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:04:22 +0000 (0:00:06.833)       0:02:40.920 ********** ",
                                "=============================================================================== ",
                                "designate : Copying over designate.conf -------------------------------- 17.56s",
                                "designate : Running Designate bootstrap container ---------------------- 14.15s",
                                "designate : Restart designate-api container ---------------------------- 11.14s",
                                "designate : Restart designate-backend-bind9 container ------------------- 8.50s",
                                "designate : Restart designate-mdns container ---------------------------- 8.11s",
                                "designate : Non-destructive DNS pools update ---------------------------- 6.83s",
                                "designate : Copying over config.json files for services ----------------- 6.54s",
                                "designate : Copying over pools.yaml ------------------------------------- 6.49s",
                                "service-ks-register : designate | Creating/deleting endpoints ----------- 6.30s",
                                "service-check-containers : designate | Check containers ----------------- 6.10s",
                                "service-cert-copy : designate | Copying over extra CA certificates ------ 6.05s",
                                "designate : Restart designate-central container ------------------------- 5.27s",
                                "designate : Restart designate-worker container -------------------------- 5.18s",
                                "service-ks-register : designate | Creating/deleting services ------------ 4.70s",
                                "designate : Restart designate-producer container ------------------------ 4.62s",
                                "service-ks-register : designate | Creating users ------------------------ 4.42s",
                                "service-ks-register : designate | Granting/revoking user roles ---------- 3.85s",
                                "designate : Copying over rndc.conf -------------------------------------- 3.82s",
                                "service-ks-register : designate | Creating projects --------------------- 3.53s",
                                "service-ks-register : designate | Creating roles ------------------------ 3.37s",
                                "\u001b[32m2026-03-13 01:04:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ddd6ebc-2d8f-4db1-9b18-f60ec4a6150c is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ec3768cd-3896-4b67-b736-5d828a5d96dd is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:04:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f0e57a61-725d-4e4b-97d1-231c0ed3cb56 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:04:22 +0000 (0:00:00.128)       0:00:00.128 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:04:22 +0000 (0:00:00.223)       0:00:00.351 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Keystone service] *******************************************",
                                "",
                                "TASK [Waiting for Keystone public port to be UP] *******************************",
                                "Friday 13 March 2026  01:04:22 +0000 (0:00:00.614)       0:00:00.966 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:04:23 +0000 (0:00:00.665)       0:00:01.631 ********** ",
                                "=============================================================================== ",
                                "Waiting for Keystone public port to be UP ------------------------------- 0.67s",
                                "Group hosts based on enabled services ----------------------------------- 0.62s",
                                "Group hosts based on Kolla action --------------------------------------- 0.22s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:04:27 +0000 (0:00:00.247)       0:00:00.247 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:04:28 +0000 (0:00:00.728)       0:00:00.976 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-rgw : include_tasks] ************************************************",
                                "Friday 13 March 2026  01:04:29 +0000 (0:00:00.871)       0:00:01.848 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating/deleting services] *************",
                                "Friday 13 March 2026  01:04:31 +0000 (0:00:02.109)       0:00:03.957 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating/deleting endpoints] ************",
                                "Friday 13 March 2026  01:04:34 +0000 (0:00:03.136)       0:00:07.098 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating projects] **********************",
                                "Friday 13 March 2026  01:04:40 +0000 (0:00:06.127)       0:00:13.225 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating users] *************************",
                                "Friday 13 March 2026  01:04:43 +0000 (0:00:02.786)       0:00:16.012 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating roles] *************************",
                                "Friday 13 March 2026  01:04:46 +0000 (0:00:03.058)       0:00:19.070 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Granting/revoking user roles] ***********",
                                "Friday 13 March 2026  01:04:52 +0000 (0:00:05.394)       0:00:24.465 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:04:55 +0000 (0:00:03.769)       0:00:28.234 ********** ",
                                "=============================================================================== ",
                                "service-ks-register : ceph-rgw | Creating/deleting endpoints ------------ 6.13s",
                                "service-ks-register : ceph-rgw | Creating roles ------------------------- 5.39s",
                                "service-ks-register : ceph-rgw | Granting/revoking user roles ----------- 3.77s",
                                "service-ks-register : ceph-rgw | Creating/deleting services ------------- 3.14s",
                                "service-ks-register : ceph-rgw | Creating users ------------------------- 3.06s",
                                "service-ks-register : ceph-rgw | Creating projects ---------------------- 2.79s",
                                "ceph-rgw : include_tasks ------------------------------------------------ 2.11s",
                                "Group hosts based on enabled services ----------------------------------- 0.87s",
                                "Group hosts based on Kolla action --------------------------------------- 0.73s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:03:40 +0000 (0:00:00.484)       0:00:00.484 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:03:41 +0000 (0:00:00.359)       0:00:00.843 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m",
                                "",
                                "PLAY [Apply role magnum] *******************************************************",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:03:41 +0000 (0:00:00.723)       0:00:01.567 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating/deleting services] ***************",
                                "Friday 13 March 2026  01:03:42 +0000 (0:00:01.004)       0:00:02.572 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating/deleting endpoints] **************",
                                "Friday 13 March 2026  01:03:46 +0000 (0:00:03.435)       0:00:06.007 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating projects] ************************",
                                "Friday 13 March 2026  01:03:53 +0000 (0:00:06.894)       0:00:12.905 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating users] ***************************",
                                "Friday 13 March 2026  01:03:56 +0000 (0:00:03.285)       0:00:16.190 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating roles] ***************************",
                                "Friday 13 March 2026  01:04:00 +0000 (0:00:03.789)       0:00:19.980 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Granting/revoking user roles] *************",
                                "Friday 13 March 2026  01:04:03 +0000 (0:00:03.361)       0:00:23.342 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee domain] *********************************",
                                "Friday 13 March 2026  01:04:07 +0000 (0:00:03.680)       0:00:27.023 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user] ***********************************",
                                "Friday 13 March 2026  01:04:10 +0000 (0:00:03.025)       0:00:30.048 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user role] ******************************",
                                "Friday 13 March 2026  01:04:13 +0000 (0:00:03.497)       0:00:33.545 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  01:04:16 +0000 (0:00:03.183)       0:00:36.729 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Check if policies shall be overwritten] *************************",
                                "Friday 13 March 2026  01:04:18 +0000 (0:00:01.208)       0:00:37.937 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum policy file] *****************************************",
                                "Friday 13 March 2026  01:04:18 +0000 (0:00:00.141)       0:00:38.079 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check if kubeconfig file is supplied] ***************************",
                                "Friday 13 March 2026  01:04:18 +0000 (0:00:00.400)       0:00:38.479 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over kubeconfig file] ***********************************",
                                "Friday 13 March 2026  01:04:19 +0000 (0:00:00.803)       0:00:39.283 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum kubeconfig file's path] ******************************",
                                "Friday 13 March 2026  01:04:21 +0000 (0:00:02.201)       0:00:41.485 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:04:21 +0000 (0:00:00.271)       0:00:41.756 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over extra CA certificates] *********",
                                "Friday 13 March 2026  01:04:22 +0000 (0:00:00.624)       0:00:42.381 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:04:24 +0000 (0:00:02.065)       0:00:44.446 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******",
                                "Friday 13 March 2026  01:04:25 +0000 (0:00:00.760)       0:00:45.207 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  01:04:26 +0000 (0:00:01.282)       0:00:46.490 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over magnum.conf] ***************************************",
                                "Friday 13 March 2026  01:04:28 +0000 (0:00:01.972)       0:00:48.462 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over existing policy file] ******************************",
                                "Friday 13 March 2026  01:04:34 +0000 (0:00:05.452)       0:00:53.915 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : magnum | Check containers] ********************",
                                "Friday 13 March 2026  01:04:35 +0000 (0:00:01.101)       0:00:55.016 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : magnum | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:04:39 +0000 (0:00:03.831)       0:00:58.848 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:04:39 +0000 (0:00:00.667)       0:00:59.515 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2025.1', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:04:41 +0000 (0:00:01.815)       0:01:01.330 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database] ***************************************",
                                "Friday 13 March 2026  01:04:42 +0000 (0:00:00.815)       0:01:02.146 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database user and setting permissions] **********",
                                "Friday 13 March 2026  01:04:44 +0000 (0:00:02.072)       0:01:04.218 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Running Magnum bootstrap container] *****************************",
                                "Friday 13 March 2026  01:04:46 +0000 (0:00:01.997)       0:01:06.216 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:15.374)       0:01:21.590 ********** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:00.061)       0:01:21.652 ********** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:00.064)       0:01:21.716 ********** ",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-api container] ************************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:00.065)       0:01:21.782 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-conductor container] ******************",
                                "Friday 13 March 2026  01:05:14 +0000 (0:00:12.828)       0:01:34.610 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:05:28 +0000 (0:00:13.867)       0:01:48.477 ********** ",
                                "=============================================================================== ",
                                "magnum : Running Magnum bootstrap container ---------------------------- 15.38s",
                                "magnum : Restart magnum-conductor container ---------------------------- 13.87s",
                                "magnum : Restart magnum-api container ---------------------------------- 12.83s",
                                "service-ks-register : magnum | Creating/deleting endpoints -------------- 6.90s",
                                "magnum : Copying over magnum.conf --------------------------------------- 5.45s",
                                "service-check-containers : magnum | Check containers -------------------- 3.83s",
                                "service-ks-register : magnum | Creating users --------------------------- 3.79s",
                                "service-ks-register : magnum | Granting/revoking user roles ------------- 3.68s",
                                "magnum : Creating Magnum trustee user ----------------------------------- 3.50s",
                                "service-ks-register : magnum | Creating/deleting services --------------- 3.44s",
                                "service-ks-register : magnum | Creating roles --------------------------- 3.36s",
                                "service-ks-register : magnum | Creating projects ------------------------ 3.29s",
                                "magnum : Creating Magnum trustee user role ------------------------------ 3.18s",
                                "magnum : Creating Magnum trustee domain --------------------------------- 3.03s",
                                "magnum : Copying over kubeconfig file ----------------------------------- 2.20s",
                                "magnum : Creating Magnum database --------------------------------------- 2.07s",
                                "service-cert-copy : magnum | Copying over extra CA certificates --------- 2.07s",
                                "magnum : Creating Magnum database user and setting permissions ---------- 2.00s",
                                "magnum : Copying over config.json files for services -------------------- 1.97s",
                                "service-check-containers : Include tasks -------------------------------- 1.81s",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee7163a7-a335-4c66-afc2-72b2f39b0021 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:01:41 +0000 (0:00:00.301)       0:00:00.301 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:01:42 +0000 (0:00:00.890)       0:00:01.191 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m",
                                "",
                                "PLAY [Apply role neutron] ******************************************************",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:01:43 +0000 (0:00:00.909)       0:00:02.101 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [neutron : Get container facts] *******************************************",
                                "Friday 13 March 2026  01:01:44 +0000 (0:00:00.902)       0:00:03.004 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Get container volume facts] ************************************",
                                "Friday 13 March 2026  01:01:45 +0000 (0:00:01.461)       0:00:04.465 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVN presence] ************************************",
                                "Friday 13 March 2026  01:01:46 +0000 (0:00:01.188)       0:00:05.654 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVS presence] ************************************",
                                "Friday 13 March 2026  01:01:47 +0000 (0:00:00.583)       0:00:06.237 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating/deleting services] **************",
                                "Friday 13 March 2026  01:01:47 +0000 (0:00:00.513)       0:00:06.751 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating/deleting endpoints] *************",
                                "Friday 13 March 2026  01:01:51 +0000 (0:00:03.728)       0:00:10.479 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating projects] ***********************",
                                "Friday 13 March 2026  01:01:58 +0000 (0:00:06.687)       0:00:17.167 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating users] **************************",
                                "Friday 13 March 2026  01:02:02 +0000 (0:00:03.776)       0:00:20.943 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating roles] **************************",
                                "Friday 13 March 2026  01:02:06 +0000 (0:00:03.998)       0:00:24.942 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Granting/revoking user roles] ************",
                                "Friday 13 March 2026  01:02:09 +0000 (0:00:03.541)       0:00:28.484 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:02:16 +0000 (0:00:06.810)       0:00:35.294 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Load and persist kernel modules] *****************************************",
                                "Friday 13 March 2026  01:02:17 +0000 (0:00:00.725)       0:00:36.020 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check IPv6 support] ********************************************",
                                "Friday 13 March 2026  01:02:19 +0000 (0:00:01.988)       0:00:38.008 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Friday 13 March 2026  01:02:20 +0000 (0:00:01.079)       0:00:39.087 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Ensuring config directories exist] *****************************",
                                "Friday 13 March 2026  01:02:22 +0000 (0:00:02.605)       0:00:41.693 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Check if extra ml2 plugins exists] *****************************",
                                "Friday 13 March 2026  01:02:25 +0000 (0:00:03.107)       0:00:44.801 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m",
                                "\u001b[1;35ma directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:02:26 +0000 (0:00:00.880)       0:00:45.682 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  01:02:27 +0000 (0:00:01.100)       0:00:46.783 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:02:30 +0000 (0:00:03.082)       0:00:49.866 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  01:02:33 +0000 (0:00:02.769)       0:00:52.636 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Creating TLS backend PEM File] *********************************",
                                "Friday 13 March 2026  01:02:37 +0000 (0:00:03.463)       0:00:56.099 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  01:02:39 +0000 (0:00:02.201)       0:00:58.301 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Set neutron policy file] ***************************************",
                                "Friday 13 March 2026  01:02:39 +0000 (0:00:00.091)       0:00:58.393 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over existing policy file] *****************************",
                                "Friday 13 March 2026  01:02:40 +0000 (0:00:00.749)       0:00:59.143 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over config.json files for services] *******************",
                                "Friday 13 March 2026  01:02:42 +0000 (0:00:02.707)       0:01:01.851 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron.conf] *************************************",
                                "Friday 13 March 2026  01:02:46 +0000 (0:00:03.832)       0:01:05.684 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_vpnaas.conf] ******************************",
                                "Friday 13 March 2026  01:02:53 +0000 (0:00:06.656)       0:01:12.340 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ssh key] ******************************************",
                                "Friday 13 March 2026  01:02:56 +0000 (0:00:02.999)       0:01:15.340 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ml2_conf.ini] *************************************",
                                "Friday 13 March 2026  01:02:59 +0000 (0:00:02.948)       0:01:18.288 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over openvswitch_agent.ini] ****************************",
                                "Friday 13 March 2026  01:03:02 +0000 (0:00:03.509)       0:01:21.798 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over sriov_agent.ini] **********************************",
                                "Friday 13 March 2026  01:03:05 +0000 (0:00:02.736)       0:01:24.535 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over mlnx_agent.ini] ***********************************",
                                "Friday 13 March 2026  01:03:08 +0000 (0:00:02.410)       0:01:26.945 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over eswitchd.conf] ************************************",
                                "Friday 13 March 2026  01:03:09 +0000 (0:00:01.812)       0:01:28.758 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dhcp_agent.ini] ***********************************",
                                "Friday 13 March 2026  01:03:11 +0000 (0:00:01.903)       0:01:30.662 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dnsmasq.conf] *************************************",
                                "Friday 13 March 2026  01:03:14 +0000 (0:00:02.371)       0:01:33.033 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over l3_agent.ini] *************************************",
                                "Friday 13 March 2026  01:03:16 +0000 (0:00:02.222)       0:01:35.256 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over fwaas_driver.ini] *********************************",
                                "Friday 13 March 2026  01:03:17 +0000 (0:00:01.614)       0:01:36.870 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metadata_agent.ini] *******************************",
                                "Friday 13 March 2026  01:03:19 +0000 (0:00:01.901)       0:01:38.772 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************",
                                "Friday 13 March 2026  01:03:21 +0000 (0:00:01.686)       0:01:40.459 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metering_agent.ini] *******************************",
                                "Friday 13 March 2026  01:03:24 +0000 (0:00:02.934)       0:01:43.393 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ironic_neutron_agent.ini] *************************",
                                "Friday 13 March 2026  01:03:26 +0000 (0:00:02.290)       0:01:45.684 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over bgp_dragent.ini] **********************************",
                                "Friday 13 March 2026  01:03:28 +0000 (0:00:01.541)       0:01:47.226 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ovn_agent.ini] ************************************",
                                "Friday 13 March 2026  01:03:30 +0000 (0:00:02.018)       0:01:49.244 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over nsx.ini] ******************************************",
                                "Friday 13 March 2026  01:03:32 +0000 (0:00:02.047)       0:01:51.292 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copy neutron-l3-agent-wrapper script] **************************",
                                "Friday 13 March 2026  01:03:34 +0000 (0:00:02.261)       0:01:53.554 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over extra ml2 plugins] ********************************",
                                "Friday 13 March 2026  01:03:37 +0000 (0:00:02.601)       0:01:56.155 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************",
                                "Friday 13 March 2026  01:03:39 +0000 (0:00:02.359)       0:01:58.514 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_taas.conf] ********************************",
                                "Friday 13 March 2026  01:03:41 +0000 (0:00:02.214)       0:02:00.729 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [service-check-containers : neutron | Check containers] *******************",
                                "Friday 13 March 2026  01:03:44 +0000 (0:00:03.103)       0:02:03.832 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : neutron | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:03:47 +0000 (0:00:02.887)       0:02:06.720 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:03:48 +0000 (0:00:00.673)       0:02:07.394 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2025.1', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:03:51 +0000 (0:00:02.683)       0:02:10.078 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database] *************************************",
                                "Friday 13 March 2026  01:03:51 +0000 (0:00:00.503)       0:02:10.581 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database user and setting permissions] ********",
                                "Friday 13 March 2026  01:03:53 +0000 (0:00:02.014)       0:02:12.595 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Running Neutron bootstrap container] ***************************",
                                "Friday 13 March 2026  01:03:56 +0000 (0:00:02.368)       0:02:14.964 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:37.479)       0:02:52.443 ********** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.051)       0:02:52.495 ********** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.161)       0:02:52.657 ********** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.048)       0:02:52.705 ********** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.051)       0:02:52.757 ********** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.047)       0:02:52.804 ********** ",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-server container] *******************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:00.048)       0:02:52.853 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******",
                                "Friday 13 March 2026  01:05:00 +0000 (0:00:26.639)       0:03:19.493 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:05:53 +0000 (0:00:52.467)       0:04:11.961 ********** ",
                                "=============================================================================== ",
                                "neutron : Restart neutron-ovn-metadata-agent container ----------------- 52.47s",
                                "neutron : Running Neutron bootstrap container -------------------------- 37.48s",
                                "neutron : Restart neutron-server container ----------------------------- 26.64s",
                                "service-ks-register : neutron | Granting/revoking user roles ------------ 6.81s",
                                "service-ks-register : neutron | Creating/deleting endpoints ------------- 6.69s",
                                "neutron : Copying over neutron.conf ------------------------------------- 6.66s",
                                "service-ks-register : neutron | Creating users -------------------------- 4.00s",
                                "neutron : Copying over config.json files for services ------------------- 3.83s",
                                "service-ks-register : neutron | Creating projects ----------------------- 3.78s",
                                "service-ks-register : neutron | Creating/deleting services -------------- 3.73s",
                                "service-ks-register : neutron | Creating roles -------------------------- 3.54s",
                                "neutron : Copying over ml2_conf.ini ------------------------------------- 3.51s",
                                "service-cert-copy : neutron | Copying over backend internal TLS key ----- 3.46s",
                                "neutron : Ensuring config directories exist ----------------------------- 3.11s",
                                "neutron : Copying over neutron_taas.conf -------------------------------- 3.10s",
                                "service-cert-copy : neutron | Copying over extra CA certificates -------- 3.08s",
                                "neutron : Copying over neutron_vpnaas.conf ------------------------------ 3.00s",
                                "neutron : Copying over ssh key ------------------------------------------ 2.95s",
                                "neutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 2.93s",
                                "service-check-containers : neutron | Check containers ------------------- 2.89s",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:05:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:06:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 58a36da8-9170-4ab5-8785-f8f896b26108 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:04:27 +0000 (0:00:00.366)       0:00:00.367 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:04:27 +0000 (0:00:00.736)       0:00:01.103 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m",
                                "",
                                "PLAY [Apply role prometheus] ***************************************************",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Friday 13 March 2026  01:04:28 +0000 (0:00:00.612)       0:00:01.715 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [prometheus : Ensuring config directories exist] **************************",
                                "Friday 13 March 2026  01:04:29 +0000 (0:00:01.540)       0:00:03.256 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Friday 13 March 2026  01:04:33 +0000 (0:00:03.150)       0:00:06.406 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****",
                                "Friday 13 March 2026  01:04:34 +0000 (0:00:01.580)       0:00:07.987 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:04:41 +0000 (0:00:07.320)       0:00:15.308 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***",
                                "Friday 13 March 2026  01:04:44 +0000 (0:00:02.668)       0:00:17.976 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over config.json files] *****************************",
                                "Friday 13 March 2026  01:04:47 +0000 (0:00:02.651)       0:00:20.628 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom prometheus alert rules files] *******************",
                                "Friday 13 March 2026  01:04:53 +0000 (0:00:06.094)       0:00:26.723 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom prometheus alert rules files] ***********",
                                "Friday 13 March 2026  01:04:54 +0000 (0:00:01.086)       0:00:27.809 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus common config overrides] ********************",
                                "Friday 13 March 2026  01:04:55 +0000 (0:00:00.604)       0:00:28.414 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus host config overrides] **********************",
                                "Friday 13 March 2026  01:04:55 +0000 (0:00:00.729)       0:00:29.143 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus config file] ************************",
                                "Friday 13 March 2026  01:04:57 +0000 (0:00:01.720)       0:00:30.864 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus web config file] ********************",
                                "Friday 13 March 2026  01:05:14 +0000 (0:00:17.252)       0:00:48.117 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus alertmanager config file] ***********",
                                "Friday 13 March 2026  01:05:18 +0000 (0:00:04.018)       0:00:52.135 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom Alertmanager alert notification templates] ******",
                                "Friday 13 March 2026  01:05:20 +0000 (0:00:01.487)       0:00:53.623 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom Alertmanager alert notification templates] ***",
                                "Friday 13 March 2026  01:05:20 +0000 (0:00:00.664)       0:00:54.287 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************",
                                "Friday 13 March 2026  01:05:21 +0000 (0:00:00.578)       0:00:54.866 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying cloud config file for openstack exporter] ***********",
                                "Friday 13 March 2026  01:05:23 +0000 (0:00:01.829)       0:00:56.696 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying config file for blackbox exporter] ******************",
                                "Friday 13 March 2026  01:05:24 +0000 (0:00:01.320)       0:00:58.016 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Find extra prometheus server config files] ******************",
                                "Friday 13 March 2026  01:05:26 +0000 (0:00:01.627)       0:00:59.644 ********** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m",
                                "\u001b[1;35mnot a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Create subdirectories for extra config files] ***************",
                                "Friday 13 March 2026  01:05:27 +0000 (0:00:01.084)       0:01:00.729 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Template extra prometheus server config files] **************",
                                "Friday 13 March 2026  01:05:28 +0000 (0:00:00.822)       0:01:01.551 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-check-containers : prometheus | Check containers] ****************",
                                "Friday 13 March 2026  01:05:28 +0000 (0:00:00.595)       0:01:02.146 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : prometheus | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:05:33 +0000 (0:00:04.718)       0:01:06.865 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:05:34 +0000 (0:00:00.725)       0:01:07.591 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-server:2025.1', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_server:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True, 'backend_http_extra': ['option httpchk GET /-/ready HTTP/1.0', \"http-check send hdr Authorization 'Basic aGFwcm94eTptdWVNaWV4aWUzYW5nb28wZnVjaGFod2VlUXVhaEpvbw=='\"]}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2025.1', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True, 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'cap_add': ['CAP_NET_RAW'], 'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Creating prometheus database user and setting permissions] ***",
                                "Friday 13 March 2026  01:05:35 +0000 (0:00:01.748)       0:01:09.339 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:01.033)       0:01:10.372 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.051)       0:01:10.424 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.047)       0:01:10.472 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.048)       0:01:10.521 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.054)       0:01:10.575 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.049)       0:01:10.624 ********** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.173)       0:01:10.798 ********** ",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-server container] *************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.064)       0:01:10.863 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******",
                                "Friday 13 March 2026  01:05:57 +0000 (0:00:20.102)       0:01:30.966 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****",
                                "Friday 13 March 2026  01:06:11 +0000 (0:00:13.596)       0:01:44.562 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***",
                                "Friday 13 March 2026  01:06:17 +0000 (0:00:05.864)       0:01:50.426 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********",
                                "Friday 13 March 2026  01:06:23 +0000 (0:00:06.180)       0:01:56.606 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******",
                                "Friday 13 March 2026  01:06:36 +0000 (0:00:13.424)       0:02:10.030 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***",
                                "Friday 13 March 2026  01:06:43 +0000 (0:00:06.849)       0:02:16.880 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***",
                                "Friday 13 March 2026  01:06:54 +0000 (0:00:10.572)       0:02:27.453 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***",
                                "Friday 13 March 2026  01:06:59 +0000 (0:00:05.117)       0:02:32.570 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=10  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:07:09 +0000 (0:00:09.842)       0:02:42.413 ********** ",
                                "=============================================================================== ",
                                "prometheus : Restart prometheus-server container ----------------------- 20.10s",
                                "prometheus : Copying over prometheus config file ----------------------- 17.25s",
                                "prometheus : Restart prometheus-node-exporter container ---------------- 13.60s",
                                "prometheus : Restart prometheus-cadvisor container --------------------- 13.42s",
                                "prometheus : Restart prometheus-elasticsearch-exporter container ------- 10.57s",
                                "prometheus : Restart prometheus-libvirt-exporter container -------------- 9.84s",
                                "service-cert-copy : prometheus | Copying over extra CA certificates ----- 7.32s",
                                "prometheus : Restart prometheus-alertmanager container ------------------ 6.85s",
                                "prometheus : Restart prometheus-memcached-exporter container ------------ 6.18s",
                                "prometheus : Copying over config.json files ----------------------------- 6.09s",
                                "prometheus : Restart prometheus-mysqld-exporter container --------------- 5.86s",
                                "prometheus : Restart prometheus-blackbox-exporter container ------------- 5.12s",
                                "service-check-containers : prometheus | Check containers ---------------- 4.72s",
                                "prometheus : Copying over prometheus web config file -------------------- 4.02s",
                                "prometheus : Ensuring config directories exist -------------------------- 3.15s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS certificate --- 2.67s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS key --- 2.65s",
                                "prometheus : Copying over my.cnf for mysqld_exporter -------------------- 1.83s",
                                "service-check-containers : Include tasks -------------------------------- 1.75s",
                                "prometheus : Find prometheus host config overrides ---------------------- 1.72s",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2f98c276-eca3-477f-9d0c-64b60c3e5154 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:05:00 +0000 (0:00:00.237)       0:00:00.237 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:05:00 +0000 (0:00:00.317)       0:00:00.555 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m",
                                "",
                                "PLAY [Apply role glance] *******************************************************",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:00.430)       0:00:00.985 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating/deleting services] ***************",
                                "Friday 13 March 2026  01:05:01 +0000 (0:00:00.846)       0:00:01.835 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating/deleting endpoints] **************",
                                "Friday 13 March 2026  01:05:05 +0000 (0:00:03.362)       0:00:05.197 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating projects] ************************",
                                "Friday 13 March 2026  01:05:11 +0000 (0:00:06.279)       0:00:11.477 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating users] ***************************",
                                "Friday 13 March 2026  01:05:15 +0000 (0:00:03.457)       0:00:14.935 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating roles] ***************************",
                                "Friday 13 March 2026  01:05:18 +0000 (0:00:03.576)       0:00:18.511 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Granting/revoking user roles] *************",
                                "Friday 13 March 2026  01:05:21 +0000 (0:00:03.140)       0:00:21.652 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  01:05:25 +0000 (0:00:03.622)       0:00:25.274 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:05:28 +0000 (0:00:03.017)       0:00:28.292 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [glance : Ensuring glance service ceph config subdir exists] **************",
                                "Friday 13 March 2026  01:05:29 +0000 (0:00:00.600)       0:00:28.892 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copy over multiple ceph configs for Glance] *********************",
                                "Friday 13 March 2026  01:05:33 +0000 (0:00:04.741)       0:00:33.634 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Copy over ceph Glance keyrings] *********************************",
                                "Friday 13 March 2026  01:05:35 +0000 (0:00:01.639)       0:00:35.273 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'pool': 'images', 'user': 'glance', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directory has correct owner and permission] *****",
                                "Friday 13 March 2026  01:05:36 +0000 (0:00:01.136)       0:00:36.409 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check if policies shall be overwritten] *************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.704)       0:00:37.114 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Set glance policy file] *****************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.119)       0:00:37.234 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:05:37 +0000 (0:00:00.260)       0:00:37.494 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over extra CA certificates] *********",
                                "Friday 13 March 2026  01:05:38 +0000 (0:00:00.460)       0:00:37.955 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:05:41 +0000 (0:00:03.491)       0:00:41.447 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS key] ******",
                                "Friday 13 March 2026  01:05:44 +0000 (0:00:02.705)       0:00:44.152 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating TLS backend PEM File] **********************************",
                                "Friday 13 March 2026  01:05:47 +0000 (0:00:03.115)       0:00:47.268 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  01:05:50 +0000 (0:00:03.419)       0:00:50.688 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-api.conf] ***********************************",
                                "Friday 13 March 2026  01:05:55 +0000 (0:00:04.447)       0:00:55.136 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-cache.conf for glance_api] ******************",
                                "Friday 13 March 2026  01:06:03 +0000 (0:00:08.460)       0:01:03.596 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-image-import.conf] **************************",
                                "Friday 13 March 2026  01:06:06 +0000 (0:00:03.144)       0:01:06.740 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over property-protections-rules.conf] *******************",
                                "Friday 13 March 2026  01:06:09 +0000 (0:00:02.564)       0:01:09.305 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over existing policy file] ******************************",
                                "Friday 13 March 2026  01:06:13 +0000 (0:00:03.904)       0:01:13.210 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-haproxy-tls.cfg] ****************************",
                                "Friday 13 March 2026  01:06:13 +0000 (0:00:00.286)       0:01:13.496 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostnqn' file for glance_api] ***********************",
                                "Friday 13 March 2026  01:06:17 +0000 (0:00:03.545)       0:01:17.042 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-check-containers : glance | Check containers] ********************",
                                "Friday 13 March 2026  01:06:22 +0000 (0:00:05.006)       0:01:22.049 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : glance | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:06:28 +0000 (0:00:06.367)       0:01:28.416 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:06:28 +0000 (0:00:00.343)       0:01:28.759 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2025.1', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h', 'option httpchk'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:06:32 +0000 (0:00:03.436)       0:01:32.195 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database] ***************************************",
                                "Friday 13 March 2026  01:06:32 +0000 (0:00:00.368)       0:01:32.564 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database user and setting permissions] **********",
                                "Friday 13 March 2026  01:06:34 +0000 (0:00:01.697)       0:01:34.262 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Enable log_bin_trust_function_creators function] ****************",
                                "Friday 13 March 2026  01:06:36 +0000 (0:00:01.830)       0:01:36.093 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Running Glance bootstrap container] *****************************",
                                "Friday 13 March 2026  01:06:38 +0000 (0:00:02.016)       0:01:38.109 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Disable log_bin_trust_function_creators function] ***************",
                                "Friday 13 March 2026  01:07:04 +0000 (0:00:25.870)       0:02:03.980 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:06 +0000 (0:00:01.956)       0:02:05.936 ********** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:06 +0000 (0:00:00.059)       0:02:05.995 ********** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:06 +0000 (0:00:00.065)       0:02:06.061 ********** ",
                                "",
                                "RUNNING HANDLER [glance : Restart glance-api container] ************************",
                                "Friday 13 March 2026  01:07:06 +0000 (0:00:00.079)       0:02:06.140 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=28  \u001b[0m \u001b[0;33mchanged=20  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=17  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:07:31 +0000 (0:00:24.834)       0:02:30.975 ********** ",
                                "=============================================================================== ",
                                "glance : Running Glance bootstrap container ---------------------------- 25.87s",
                                "glance : Restart glance-api container ---------------------------------- 24.83s",
                                "glance : Copying over glance-api.conf ----------------------------------- 8.46s",
                                "service-check-containers : glance | Check containers -------------------- 6.37s",
                                "service-ks-register : glance | Creating/deleting endpoints -------------- 6.28s",
                                "glance : Generating 'hostnqn' file for glance_api ----------------------- 5.01s",
                                "glance : Ensuring glance service ceph config subdir exists -------------- 4.74s",
                                "glance : Copying over config.json files for services -------------------- 4.45s",
                                "glance : Copying over property-protections-rules.conf ------------------- 3.90s",
                                "service-ks-register : glance | Granting/revoking user roles ------------- 3.62s",
                                "service-ks-register : glance | Creating users --------------------------- 3.58s",
                                "glance : Copying over glance-haproxy-tls.cfg ---------------------------- 3.55s",
                                "service-cert-copy : glance | Copying over extra CA certificates --------- 3.49s",
                                "service-ks-register : glance | Creating projects ------------------------ 3.46s",
                                "service-check-containers : Include tasks -------------------------------- 3.44s",
                                "glance : Creating TLS backend PEM File ---------------------------------- 3.42s",
                                "service-ks-register : glance | Creating/deleting services --------------- 3.36s",
                                "glance : Copying over glance-cache.conf for glance_api ------------------ 3.14s",
                                "service-ks-register : glance | Creating roles --------------------------- 3.14s",
                                "service-cert-copy : glance | Copying over backend internal TLS key ------ 3.12s",
                                "\u001b[32m2026-03-13 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:07:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ccf61df7-9839-4a64-be2e-f28b11da3826 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:05:34 +0000 (0:00:00.332)       0:00:00.332 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:05:35 +0000 (0:00:00.472)       0:00:00.805 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m",
                                "",
                                "PLAY [Apply role cinder] *******************************************************",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:05:35 +0000 (0:00:00.331)       0:00:01.137 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating/deleting services] ***************",
                                "Friday 13 March 2026  01:05:35 +0000 (0:00:00.547)       0:00:01.684 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder (block-storage))\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating/deleting endpoints] **************",
                                "Friday 13 March 2026  01:05:41 +0000 (0:00:05.941)       0:00:07.625 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> https://api-int.testbed.osism.xyz:8776/v3 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> https://api.testbed.osism.xyz:8776/v3 -> public)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating projects] ************************",
                                "Friday 13 March 2026  01:05:54 +0000 (0:00:12.584)       0:00:20.210 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating users] ***************************",
                                "Friday 13 March 2026  01:05:57 +0000 (0:00:03.268)       0:00:23.478 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating roles] ***************************",
                                "Friday 13 March 2026  01:06:01 +0000 (0:00:03.499)       0:00:26.977 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Granting/revoking user roles] *************",
                                "Friday 13 March 2026  01:06:04 +0000 (0:00:03.438)       0:00:30.416 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directories exist] ******************************",
                                "Friday 13 March 2026  01:06:11 +0000 (0:00:06.516)       0:00:36.932 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:06:13 +0000 (0:00:02.596)       0:00:39.529 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:06:14 +0000 (0:00:00.354)       0:00:39.883 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring cinder service ceph config subdirs exists] *************",
                                "Friday 13 March 2026  01:06:14 +0000 (0:00:00.762)       0:00:40.645 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over multiple ceph.conf for cinder services] ************",
                                "Friday 13 March 2026  01:06:16 +0000 (0:00:01.866)       0:00:42.512 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True}])\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************",
                                "Friday 13 March 2026  01:06:22 +0000 (0:00:05.972)       0:00:48.485 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************",
                                "Friday 13 March 2026  01:06:25 +0000 (0:00:02.868)       0:00:51.354 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder', 'pool': 'volumes', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'backend_name': 'rbd-1', 'cluster': 'ceph', 'user': 'cinder-backup', 'pool': 'backups', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directory has correct owner and permission] *****",
                                "Friday 13 March 2026  01:06:29 +0000 (0:00:03.729)       0:00:55.083 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Check if policies shall be overwritten] *************************",
                                "Friday 13 March 2026  01:06:30 +0000 (0:00:01.206)       0:00:56.290 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Set cinder policy file] *****************************************",
                                "Friday 13 March 2026  01:06:30 +0000 (0:00:00.120)       0:00:56.410 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:06:31 +0000 (0:00:00.331)       0:00:56.741 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over extra CA certificates] *********",
                                "Friday 13 March 2026  01:06:31 +0000 (0:00:00.713)       0:00:57.454 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:06:34 +0000 (0:00:03.162)       0:01:00.617 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******",
                                "Friday 13 March 2026  01:06:35 +0000 (0:00:00.669)       0:01:01.286 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over config.json files for services] ********************",
                                "Friday 13 March 2026  01:06:36 +0000 (0:00:01.233)       0:01:02.520 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder-wsgi.conf] **********************************",
                                "Friday 13 March 2026  01:06:40 +0000 (0:00:03.918)       0:01:06.438 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Configure uWSGI for Cinder] **********************************************",
                                "Friday 13 March 2026  01:06:41 +0000 (0:00:00.714)       0:01:07.153 ********** ",
                                "\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-uwsgi-config : Copying over cinder-api uWSGI config] *************",
                                "Friday 13 March 2026  01:06:42 +0000 (0:00:01.436)       0:01:08.590 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder.conf] ***************************************",
                                "Friday 13 March 2026  01:06:45 +0000 (0:00:02.748)       0:01:11.339 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************",
                                "Friday 13 March 2026  01:06:55 +0000 (0:00:10.013)       0:01:21.352 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over existing policy file] ******************************",
                                "Friday 13 March 2026  01:06:57 +0000 (0:00:01.612)       0:01:22.965 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over nfs_shares files for cinder_volume] ****************",
                                "Friday 13 March 2026  01:06:57 +0000 (0:00:00.673)       0:01:23.638 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : cinder | Check containers] ********************",
                                "Friday 13 March 2026  01:06:58 +0000 (0:00:00.344)       0:01:23.982 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : cinder | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:07:01 +0000 (0:00:03.045)       0:01:27.028 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:07:01 +0000 (0:00:00.564)       0:01:27.592 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2025.1', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'wsgi': 'cinder.wsgi.api:application', 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2025.1', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2025.1', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2025.1', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Friday 13 March 2026  01:07:02 +0000 (0:00:00.875)       0:01:28.467 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database] ***************************************",
                                "Friday 13 March 2026  01:07:03 +0000 (0:00:00.293)       0:01:28.761 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database user and setting permissions] **********",
                                "Friday 13 March 2026  01:07:04 +0000 (0:00:01.804)       0:01:30.565 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Running Cinder bootstrap container] *****************************",
                                "Friday 13 March 2026  01:07:07 +0000 (0:00:02.545)       0:01:33.111 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:25 +0000 (0:00:18.031)       0:01:51.143 ********** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:25 +0000 (0:00:00.066)       0:01:51.210 ********** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Friday 13 March 2026  01:07:25 +0000 (0:00:00.078)       0:01:51.289 ********** ",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-api container] ************************",
                                "Friday 13 March 2026  01:07:25 +0000 (0:00:00.066)       0:01:51.355 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************",
                                "Friday 13 March 2026  01:07:49 +0000 (0:00:23.444)       0:02:14.799 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-volume container] *********************",
                                "Friday 13 March 2026  01:07:54 +0000 (0:00:05.290)       0:02:20.089 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-backup container] *********************",
                                "Friday 13 March 2026  01:08:21 +0000 (0:00:27.364)       0:02:47.454 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***",
                                "Friday 13 March 2026  01:08:32 +0000 (0:00:10.790)       0:02:58.244 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:08:32 +0000 (0:00:00.295)       0:02:58.540 ********** ",
                                "=============================================================================== ",
                                "cinder : Restart cinder-volume container ------------------------------- 27.36s",
                                "cinder : Restart cinder-api container ---------------------------------- 23.44s",
                                "cinder : Running Cinder bootstrap container ---------------------------- 18.03s",
                                "service-ks-register : cinder | Creating/deleting endpoints ------------- 12.58s",
                                "cinder : Restart cinder-backup container ------------------------------- 10.79s",
                                "cinder : Copying over cinder.conf -------------------------------------- 10.01s",
                                "service-ks-register : cinder | Granting/revoking user roles ------------- 6.52s",
                                "cinder : Copying over multiple ceph.conf for cinder services ------------ 5.97s",
                                "service-ks-register : cinder | Creating/deleting services --------------- 5.94s",
                                "cinder : Restart cinder-scheduler container ----------------------------- 5.29s",
                                "cinder : Copying over config.json files for services -------------------- 3.92s",
                                "cinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.73s",
                                "service-ks-register : cinder | Creating users --------------------------- 3.50s",
                                "service-ks-register : cinder | Creating roles --------------------------- 3.44s",
                                "service-ks-register : cinder | Creating projects ------------------------ 3.27s",
                                "service-cert-copy : cinder | Copying over extra CA certificates --------- 3.16s",
                                "service-check-containers : cinder | Check containers -------------------- 3.05s",
                                "cinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.87s",
                                "service-uwsgi-config : Copying over cinder-api uWSGI config ------------- 2.75s",
                                "cinder : Ensuring config directories exist ------------------------------ 2.60s",
                                "\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:08:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 75a3d637-8802-4b75-9a87-55aad268e119 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:07:35 +0000 (0:00:00.254)       0:00:00.255 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:07:35 +0000 (0:00:00.297)       0:00:00.552 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m",
                                "",
                                "PLAY [Apply role grafana] ******************************************************",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:07:36 +0000 (0:00:00.414)       0:00:00.967 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [grafana : Ensuring config directories exist] *****************************",
                                "Friday 13 March 2026  01:07:36 +0000 (0:00:00.487)       0:00:01.454 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Check if extra configuration file exists] **********************",
                                "Friday 13 March 2026  01:07:37 +0000 (0:00:00.665)       0:00:02.119 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:07:38 +0000 (0:00:00.884)       0:00:03.004 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  01:07:38 +0000 (0:00:00.647)       0:00:03.652 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:07:40 +0000 (0:00:01.267)       0:00:04.920 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  01:07:40 +0000 (0:00:00.490)       0:00:05.411 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over config.json files] ********************************",
                                "Friday 13 March 2026  01:07:41 +0000 (0:00:00.874)       0:00:06.286 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over grafana.ini] **************************************",
                                "Friday 13 March 2026  01:07:42 +0000 (0:00:01.145)       0:00:07.431 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over extra configuration file] *************************",
                                "Friday 13 March 2026  01:07:44 +0000 (0:00:01.230)       0:00:08.662 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Configuring Prometheus as data source for Grafana] *************",
                                "Friday 13 March 2026  01:07:44 +0000 (0:00:00.458)       0:00:09.120 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "",
                                "TASK [grafana : Configuring dashboards provisioning] ***************************",
                                "Friday 13 March 2026  01:07:45 +0000 (0:00:01.181)       0:00:10.302 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "",
                                "TASK [grafana : Check if the folder for custom grafana dashboards exists] ******",
                                "Friday 13 March 2026  01:07:46 +0000 (0:00:01.104)       0:00:11.406 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : Remove templated Grafana dashboards] ***************************",
                                "Friday 13 March 2026  01:07:47 +0000 (0:00:00.746)       0:00:12.152 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over custom dashboards] ********************************",
                                "Friday 13 March 2026  01:07:48 +0000 (0:00:00.692)       0:00:12.845 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : grafana | Check containers] *******************",
                                "Friday 13 March 2026  01:07:49 +0000 (0:00:01.458)       0:00:14.303 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : grafana | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:07:50 +0000 (0:00:01.342)       0:00:15.646 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:07:51 +0000 (0:00:00.363)       0:00:16.009 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2025.1', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database] *************************************",
                                "Friday 13 March 2026  01:07:52 +0000 (0:00:00.709)       0:00:16.719 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database user and setting permissions] ********",
                                "Friday 13 March 2026  01:07:54 +0000 (0:00:02.253)       0:00:18.973 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:07:56 +0000 (0:00:02.344)       0:00:21.317 ********** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.381 ********** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.444 ********** ",
                                "",
                                "RUNNING HANDLER [grafana : Restart first grafana container] ********************",
                                "Friday 13 March 2026  01:07:56 +0000 (0:00:00.063)       0:00:21.507 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********",
                                "Friday 13 March 2026  01:08:03 +0000 (0:00:07.102)       0:00:28.610 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Restart remaining grafana containers] ***************",
                                "Friday 13 March 2026  01:08:17 +0000 (0:00:13.893)       0:00:42.503 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Wait for grafana application ready] ****************************",
                                "Friday 13 March 2026  01:08:52 +0000 (0:00:34.933)       0:01:17.437 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Remove old grafana docker volume] ******************************",
                                "Friday 13 March 2026  01:08:54 +0000 (0:00:01.878)       0:01:19.316 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Enable grafana datasources] ************************************",
                                "Friday 13 March 2026  01:08:54 +0000 (0:00:00.279)       0:01:19.595 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Disable Getting Started panel] *********************************",
                                "Friday 13 March 2026  01:08:56 +0000 (0:00:01.982)       0:01:21.577 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=13  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:08:57 +0000 (0:00:00.250)       0:01:21.828 ********** ",
                                "=============================================================================== ",
                                "grafana : Restart remaining grafana containers ------------------------- 34.93s",
                                "grafana : Waiting for grafana to start on first node ------------------- 13.89s",
                                "grafana : Restart first grafana container ------------------------------- 7.10s",
                                "grafana : Creating grafana database user and setting permissions -------- 2.34s",
                                "grafana : Creating grafana database ------------------------------------- 2.25s",
                                "grafana : Enable grafana datasources ------------------------------------ 1.98s",
                                "grafana : Wait for grafana application ready ---------------------------- 1.88s",
                                "grafana : Copying over custom dashboards -------------------------------- 1.46s",
                                "service-check-containers : grafana | Check containers ------------------- 1.34s",
                                "service-cert-copy : grafana | Copying over extra CA certificates -------- 1.27s",
                                "grafana : Copying over grafana.ini -------------------------------------- 1.23s",
                                "grafana : Configuring Prometheus as data source for Grafana ------------- 1.18s",
                                "grafana : Copying over config.json files -------------------------------- 1.15s",
                                "grafana : Configuring dashboards provisioning --------------------------- 1.10s",
                                "grafana : Check if extra configuration file exists ---------------------- 0.88s",
                                "service-cert-copy : grafana | Copying over backend internal TLS key ----- 0.87s",
                                "grafana : Check if the folder for custom grafana dashboards exists ------ 0.75s",
                                "service-check-containers : Include tasks -------------------------------- 0.71s",
                                "grafana : Remove templated Grafana dashboards --------------------------- 0.69s",
                                "grafana : Ensuring config directories exist ----------------------------- 0.67s",
                                "\u001b[32m2026-03-13 01:09:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:09:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a7c0a400-6678-4bf6-be30-b73f7a2a1a64 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:10:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:11:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:12:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-03-13 01:13:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 77c6c484-bfb0-40f8-ba17-8bcab28dbc7b is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:07:12 +0000 (0:00:00.130)       0:00:00.130 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:07:13 +0000 (0:00:00.221)       0:00:00.352 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Nova service] ***********************************************",
                                "",
                                "TASK [Waiting for Nova public port to be UP] ***********************************",
                                "Friday 13 March 2026  01:07:13 +0000 (0:00:00.554)       0:00:00.907 ********** ",
                                "",
                                "STILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:10:07 +0000 (0:02:53.817)       0:02:54.724 ********** ",
                                "=============================================================================== ",
                                "Waiting for Nova public port to be UP --------------------------------- 173.82s",
                                "Group hosts based on enabled services ----------------------------------- 0.55s",
                                "Group hosts based on Kolla action --------------------------------------- 0.22s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on OpenStack release] **********************************",
                                "Friday 13 March 2026  01:05:59 +0000 (0:00:00.612)       0:00:00.612 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:06:01 +0000 (0:00:01.949)       0:00:02.561 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:06:02 +0000 (0:00:01.465)       0:00:04.027 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova API databases] ********************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Friday 13 March 2026  01:06:03 +0000 (0:00:00.706)       0:00:04.734 ********** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases] ******************************************",
                                "Friday 13 March 2026  01:06:03 +0000 (0:00:00.549)       0:00:05.283 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases user and setting permissions] *************",
                                "Friday 13 March 2026  01:06:07 +0000 (0:00:03.931)       0:00:09.215 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Friday 13 March 2026  01:06:11 +0000 (0:00:03.467)       0:00:12.682 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for nova-api-bootstrap] ************",
                                "Friday 13 March 2026  01:06:11 +0000 (0:00:00.746)       0:00:13.429 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************",
                                "Friday 13 March 2026  01:06:13 +0000 (0:00:01.675)       0:00:15.104 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Friday 13 March 2026  01:06:16 +0000 (0:00:03.180)       0:00:18.285 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Friday 13 March 2026  01:06:17 +0000 (0:00:00.430)       0:00:18.715 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Create cell0 mappings] ********************************************",
                                "Friday 13 March 2026  01:06:46 +0000 (0:00:28.997)       0:00:47.712 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Friday 13 March 2026  01:07:01 +0000 (0:00:14.834)       0:01:02.547 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Friday 13 March 2026  01:07:14 +0000 (0:00:13.131)       0:01:15.679 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Update cell0 mappings] ********************************************",
                                "Friday 13 March 2026  01:07:15 +0000 (0:00:00.904)       0:01:16.583 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Friday 13 March 2026  01:07:15 +0000 (0:00:00.456)       0:01:17.040 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Friday 13 March 2026  01:07:15 +0000 (0:00:00.393)       0:01:17.433 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Friday 13 March 2026  01:07:33 +0000 (0:00:17.826)       0:01:35.260 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova cell databases] *******************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Friday 13 March 2026  01:07:34 +0000 (0:00:00.310)       0:01:35.570 ********** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database] *********************************",
                                "Friday 13 March 2026  01:07:34 +0000 (0:00:00.560)       0:01:36.130 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database user and setting permissions] ****",
                                "Friday 13 March 2026  01:07:36 +0000 (0:00:01.929)       0:01:38.060 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Friday 13 March 2026  01:07:38 +0000 (0:00:01.833)       0:01:39.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Friday 13 March 2026  01:07:38 +0000 (0:00:00.333)       0:01:40.226 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Friday 13 March 2026  01:07:48 +0000 (0:00:09.884)       0:01:50.111 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Friday 13 March 2026  01:07:48 +0000 (0:00:00.331)       0:01:50.442 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Friday 13 March 2026  01:07:49 +0000 (0:00:00.731)       0:01:51.174 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******",
                                "Friday 13 March 2026  01:07:50 +0000 (0:00:00.886)       0:01:52.060 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************",
                                "Friday 13 March 2026  01:07:51 +0000 (0:00:00.917)       0:01:52.978 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Running Nova cell bootstrap container] ***********************",
                                "Friday 13 March 2026  01:07:53 +0000 (0:00:01.969)       0:01:54.947 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Friday 13 March 2026  01:08:14 +0000 (0:00:21.106)       0:02:16.053 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Friday 13 March 2026  01:08:25 +0000 (0:00:11.152)       0:02:27.206 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Create cell] *************************************************",
                                "Friday 13 March 2026  01:08:26 +0000 (0:00:00.898)       0:02:28.104 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Update cell] *************************************************",
                                "Friday 13 March 2026  01:08:38 +0000 (0:00:11.932)       0:02:40.036 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Friday 13 March 2026  01:08:39 +0000 (0:00:01.016)       0:02:41.053 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova] *********************************************************",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Friday 13 March 2026  01:08:40 +0000 (0:00:00.508)       0:02:41.562 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating/deleting services] *****************",
                                "Friday 13 March 2026  01:08:40 +0000 (0:00:00.531)       0:02:42.093 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating/deleting endpoints] ****************",
                                "Friday 13 March 2026  01:08:43 +0000 (0:00:03.067)       0:02:45.161 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating projects] **************************",
                                "Friday 13 March 2026  01:08:49 +0000 (0:00:05.679)       0:02:50.840 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating users] *****************************",
                                "Friday 13 March 2026  01:08:52 +0000 (0:00:02.717)       0:02:53.558 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating roles] *****************************",
                                "Friday 13 March 2026  01:08:55 +0000 (0:00:03.392)       0:02:56.950 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Granting/revoking user roles] ***************",
                                "Friday 13 March 2026  01:08:58 +0000 (0:00:02.769)       0:02:59.719 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Friday 13 March 2026  01:09:06 +0000 (0:00:08.218)       0:03:07.937 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Check if policies shall be overwritten] ***************************",
                                "Friday 13 March 2026  01:09:08 +0000 (0:00:01.715)       0:03:09.653 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Set nova policy file] *********************************************",
                                "Friday 13 March 2026  01:09:08 +0000 (0:00:00.143)       0:03:09.797 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check for vendordata file] ****************************************",
                                "Friday 13 March 2026  01:09:08 +0000 (0:00:00.470)       0:03:10.267 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova : Set vendordata file path] *****************************************",
                                "Friday 13 March 2026  01:09:09 +0000 (0:00:00.794)       0:03:11.062 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Friday 13 March 2026  01:09:09 +0000 (0:00:00.293)       0:03:11.356 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Friday 13 March 2026  01:09:10 +0000 (0:00:00.547)       0:03:11.903 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:09:13 +0000 (0:00:03.231)       0:03:15.135 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Friday 13 March 2026  01:09:14 +0000 (0:00:00.824)       0:03:15.959 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for services] **********************",
                                "Friday 13 March 2026  01:09:15 +0000 (0:00:00.916)       0:03:16.876 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf] *******************************************",
                                "Friday 13 March 2026  01:09:19 +0000 (0:00:03.913)       0:03:20.789 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over existing policy file] ********************************",
                                "Friday 13 March 2026  01:09:26 +0000 (0:00:07.532)       0:03:28.321 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-api-wsgi.conf] **********************************",
                                "Friday 13 March 2026  01:09:27 +0000 (0:00:00.714)       0:03:29.036 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-metadata-wsgi.conf] *****************************",
                                "Friday 13 March 2026  01:09:28 +0000 (0:00:00.648)       0:03:29.685 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over vendordata file for nova services] *******************",
                                "Friday 13 March 2026  01:09:29 +0000 (0:00:01.008)       0:03:30.694 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-metadata) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-metadata) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-metadata) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Configure uWSGI for Nova] ************************************************",
                                "Friday 13 March 2026  01:09:29 +0000 (0:00:00.591)       0:03:31.285 ********** ",
                                "\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova-api', 'port': '8774', 'workers': '2'})\u001b[0m",
                                "\u001b[0;36mincluded: service-uwsgi-config for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova-metadata', 'port': '8775', 'workers': '2'})\u001b[0m",
                                "",
                                "TASK [service-uwsgi-config : Copying over nova-api uWSGI config] ***************",
                                "Friday 13 March 2026  01:09:31 +0000 (0:00:01.372)       0:03:32.657 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-uwsgi-config : Copying over nova-metadata uWSGI config] **********",
                                "Friday 13 March 2026  01:09:33 +0000 (0:00:02.373)       0:03:35.030 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-check-containers : nova | Check containers] **********************",
                                "Friday 13 March 2026  01:09:35 +0000 (0:00:02.216)       0:03:37.247 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : nova | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:09:38 +0000 (0:00:02.910)       0:03:40.158 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:09:39 +0000 (0:00:00.701)       0:03:40.859 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.osapi_compute:application', 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-metadata', 'value': {'container_name': 'nova_metadata', 'group': 'nova-metadata', 'image': 'registry.osism.tech/kolla/nova-api:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-metadata/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8775 '], 'timeout': '30'}, 'wsgi': 'nova.wsgi.metadata:application', 'haproxy': {'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no', 'backend_http_extra': ['option httpchk']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Friday 13 March 2026  01:09:40 +0000 (0:00:00.932)       0:03:41.792 ********** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Friday 13 March 2026  01:09:40 +0000 (0:00:00.171)       0:03:41.964 ********** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Friday 13 March 2026  01:09:40 +0000 (0:00:00.131)       0:03:42.095 ********** ",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-scheduler container] **********************",
                                "Friday 13 March 2026  01:09:40 +0000 (0:00:00.224)       0:03:42.320 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-api container] ****************************",
                                "Friday 13 March 2026  01:09:59 +0000 (0:00:18.797)       0:04:01.118 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-metadata container] ***********************",
                                "Friday 13 March 2026  01:10:05 +0000 (0:00:05.886)       0:04:07.005 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova-cell] ****************************************************",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:10:10 +0000 (0:00:04.726)       0:04:11.731 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:10:11 +0000 (0:00:01.094)       0:04:12.825 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get new Libvirt version] *************************************",
                                "Friday 13 March 2026  01:10:11 +0000 (0:00:00.549)       0:04:13.374 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Cache new Libvirt version] ***********************************",
                                "Friday 13 March 2026  01:10:30 +0000 (0:00:18.866)       0:04:32.241 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get nova_libvirt image info] *********************************************",
                                "Friday 13 March 2026  01:10:31 +0000 (0:00:01.186)       0:04:33.427 ********** ",
                                "\u001b[0;36mincluded: service-image-info for testbed-node-3\u001b[0m",
                                "",
                                "TASK [service-image-info : community.docker.docker_image_info] *****************",
                                "Friday 13 March 2026  01:10:32 +0000 (0:00:00.654)       0:04:34.082 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [service-image-info : set_fact] *******************************************",
                                "Friday 13 March 2026  01:10:35 +0000 (0:00:02.933)       0:04:37.016 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [service-image-info : containers.podman.podman_image_info] ****************",
                                "Friday 13 March 2026  01:10:37 +0000 (0:00:01.922)       0:04:38.938 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [service-image-info : set_fact] *******************************************",
                                "Friday 13 March 2026  01:10:39 +0000 (0:00:02.002)       0:04:40.941 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get container facts] *****************************************",
                                "Friday 13 March 2026  01:10:41 +0000 (0:00:01.691)       0:04:42.633 ********** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "",
                                "TASK [nova-cell : Get current Libvirt version] *********************************",
                                "Friday 13 March 2026  01:10:49 +0000 (0:00:08.538)       0:04:51.171 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check that the new Libvirt version is >= current] ************",
                                "Friday 13 March 2026  01:10:54 +0000 (0:00:05.194)       0:04:56.366 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-3', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-4', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'result': False, 'changed': False, 'containers': {}, 'invocation': {'module_args': {'action': 'get_containers', 'container_engine': 'docker', 'name': ['nova_libvirt'], 'api_version': 'auto'}}, 'failed': False, 'item': 'testbed-node-5', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Load and persist br_netfilter module] ************************************",
                                "Friday 13 March 2026  01:10:58 +0000 (0:00:03.362)       0:04:59.728 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Friday 13 March 2026  01:10:59 +0000 (0:00:00.920)       0:05:00.649 ********** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Friday 13 March 2026  01:10:59 +0000 (0:00:00.653)       0:05:01.303 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Friday 13 March 2026  01:11:00 +0000 (0:00:01.057)       0:05:02.360 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************",
                                "Friday 13 March 2026  01:11:01 +0000 (0:00:00.649)       0:05:03.010 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "",
                                "TASK [nova-cell : Install udev kolla kvm rules] ********************************",
                                "Friday 13 March 2026  01:11:02 +0000 (0:00:00.999)       0:05:04.010 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Mask qemu-kvm service] ***************************************",
                                "Friday 13 March 2026  01:11:03 +0000 (0:00:00.948)       0:05:04.958 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Friday 13 March 2026  01:11:04 +0000 (0:00:01.462)       0:05:06.421 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:11:06 +0000 (0:00:02.022)       0:05:08.443 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Friday 13 March 2026  01:11:08 +0000 (0:00:01.118)       0:05:09.562 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:11:11 +0000 (0:00:03.132)       0:05:12.695 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Friday 13 March 2026  01:11:13 +0000 (0:00:01.947)       0:05:14.643 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:11:15 +0000 (0:00:02.146)       0:05:16.789 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova keyring file] *************************************",
                                "Friday 13 March 2026  01:11:16 +0000 (0:00:00.819)       0:05:17.608 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check cinder keyring file] ***********************************",
                                "Friday 13 March 2026  01:11:17 +0000 (0:00:01.158)       0:05:18.767 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract nova key from file] **********************************",
                                "Friday 13 March 2026  01:11:18 +0000 (0:00:00.977)       0:05:19.744 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract cinder key from file] ********************************",
                                "Friday 13 March 2026  01:11:18 +0000 (0:00:00.505)       0:05:20.250 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph nova keyring file] ****************************",
                                "Friday 13 March 2026  01:11:19 +0000 (0:00:00.504)       0:05:20.754 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph cinder keyring file] **************************",
                                "Friday 13 March 2026  01:11:20 +0000 (0:00:01.307)       0:05:22.062 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf] *****************************************",
                                "Friday 13 March 2026  01:11:21 +0000 (0:00:01.115)       0:05:23.178 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************",
                                "Friday 13 March 2026  01:11:25 +0000 (0:00:03.442)       0:05:26.620 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************",
                                "Friday 13 March 2026  01:11:25 +0000 (0:00:00.271)       0:05:26.891 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring libvirt secrets directory exists] *******************",
                                "Friday 13 March 2026  01:11:25 +0000 (0:00:00.400)       0:05:27.292 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing nova secret xml for libvirt] *************************",
                                "Friday 13 March 2026  01:11:26 +0000 (0:00:01.064)       0:05:28.356 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'ceph-ephemeral-nova', 'desc': 'Ceph Client Secret for Ephemeral Storage (Nova)', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'ceph-persistent-cinder', 'desc': 'Ceph Client Secret for Persistent Storage (Cinder)', 'enabled': 'yes'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing secrets key for libvirt] *****************************",
                                "Friday 13 March 2026  01:11:29 +0000 (0:00:02.783)       0:05:31.140 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************",
                                "Friday 13 March 2026  01:11:32 +0000 (0:00:02.902)       0:05:34.042 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check qemu wrapper file] *************************************",
                                "Friday 13 March 2026  01:11:34 +0000 (0:00:01.618)       0:05:35.661 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy qemu wrapper] *******************************************",
                                "Friday 13 March 2026  01:11:35 +0000 (0:00:00.912)       0:05:36.573 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check if policies shall be overwritten] **********************",
                                "Friday 13 March 2026  01:11:35 +0000 (0:00:00.469)       0:05:37.042 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set nova policy file] ****************************************",
                                "Friday 13 March 2026  01:11:35 +0000 (0:00:00.144)       0:05:37.187 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check for vendordata file] ***********************************",
                                "Friday 13 March 2026  01:11:36 +0000 (0:00:00.592)       0:05:37.779 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set vendordata file path] ************************************",
                                "Friday 13 March 2026  01:11:37 +0000 (0:00:00.816)       0:05:38.596 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for services] *****************",
                                "Friday 13 March 2026  01:11:37 +0000 (0:00:00.758)       0:05:39.355 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf] **************************************",
                                "Friday 13 March 2026  01:11:40 +0000 (0:00:03.152)       0:05:42.507 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over Nova compute provider config] *******************",
                                "Friday 13 March 2026  01:11:47 +0000 (0:00:06.547)       0:05:49.055 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt configuration] **************************",
                                "Friday 13 March 2026  01:11:48 +0000 (0:00:01.154)       0:05:50.209 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt TLS keys] *******************************",
                                "Friday 13 March 2026  01:11:52 +0000 (0:00:03.385)       0:05:53.595 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt SASL configuration] *********************",
                                "Friday 13 March 2026  01:11:52 +0000 (0:00:00.514)       0:05:54.110 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying files for nova-ssh] **********************************",
                                "Friday 13 March 2026  01:11:57 +0000 (0:00:04.714)       0:05:58.824 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying VMware vCenter CA file] ******************************",
                                "Friday 13 March 2026  01:12:03 +0000 (0:00:06.616)       0:06:05.440 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying 'release' file for nova_compute] *********************",
                                "Friday 13 March 2026  01:12:04 +0000 (0:00:00.730)       0:06:06.171 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************",
                                "Friday 13 March 2026  01:12:05 +0000 (0:00:00.571)       0:06:06.743 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over existing policy file] ***************************",
                                "Friday 13 March 2026  01:12:07 +0000 (0:00:02.278)       0:06:09.022 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over vendordata file to containers] ******************",
                                "Friday 13 March 2026  01:12:08 +0000 (0:00:01.443)       0:06:10.466 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-check-containers : nova_cell | Check containers] *****************",
                                "Friday 13 March 2026  01:12:09 +0000 (0:00:00.771)       0:06:11.237 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : nova_cell | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:12:12 +0000 (0:00:03.270)       0:06:14.508 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:12:13 +0000 (0:00:00.801)       0:06:15.309 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2025.1', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2025.1', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Friday 13 March 2026  01:12:15 +0000 (0:00:01.994)       0:06:17.304 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:16 +0000 (0:00:00.565)       0:06:17.870 ********** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:16 +0000 (0:00:00.125)       0:06:17.995 ********** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:16 +0000 (0:00:00.123)       0:06:18.119 ********** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:16 +0000 (0:00:00.292)       0:06:18.411 ********** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:17 +0000 (0:00:00.130)       0:06:18.541 ********** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Friday 13 March 2026  01:12:17 +0000 (0:00:00.132)       0:06:18.674 ********** ",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************",
                                "Friday 13 March 2026  01:12:17 +0000 (0:00:00.133)       0:06:18.807 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************",
                                "Friday 13 March 2026  01:12:23 +0000 (0:00:06.285)       0:06:25.092 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************",
                                "Friday 13 March 2026  01:12:35 +0000 (0:00:11.475)       0:06:36.568 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************",
                                "Friday 13 March 2026  01:12:56 +0000 (0:00:21.859)       0:06:58.427 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************",
                                "Friday 13 March 2026  01:13:29 +0000 (0:00:32.427)       0:07:30.855 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-4]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5]: Checking libvirt container is ready (10 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************",
                                "Friday 13 March 2026  01:13:35 +0000 (0:00:06.078)       0:07:36.933 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-compute container] *******************",
                                "Friday 13 March 2026  01:13:36 +0000 (0:00:00.663)       0:07:37.597 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***",
                                "Friday 13 March 2026  01:13:58 +0000 (0:00:22.835)       0:08:00.433 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Waiting for nova-compute services to register themselves] ****",
                                "Friday 13 March 2026  01:13:59 +0000 (0:00:00.153)       0:08:00.587 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if nova-compute service failed to register] *************",
                                "Friday 13 March 2026  01:14:19 +0000 (0:00:20.048)       0:08:20.635 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include discover_computes.yml] *******************************",
                                "Friday 13 March 2026  01:14:28 +0000 (0:00:09.719)       0:08:30.355 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Friday 13 March 2026  01:14:32 +0000 (0:00:03.677)       0:08:34.032 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Friday 13 March 2026  01:14:45 +0000 (0:00:12.980)       0:08:47.013 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if cell settings not found] *****************************",
                                "Friday 13 March 2026  01:14:46 +0000 (0:00:01.426)       0:08:48.440 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Discover nova hosts] *****************************************",
                                "Friday 13 March 2026  01:14:48 +0000 (0:00:01.574)       0:08:50.014 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "PLAY [Refresh nova scheduler cell cache] ***************************************",
                                "",
                                "TASK [nova : Refresh cell cache in nova scheduler] *****************************",
                                "Friday 13 March 2026  01:15:01 +0000 (0:00:12.956)       0:09:02.971 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova super conductor services] *****************************",
                                "",
                                "TASK [nova : Reload nova super conductor services to remove RPC version pin] ***",
                                "Friday 13 March 2026  01:15:02 +0000 (0:00:01.070)       0:09:04.042 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload Nova cell services] ***********************************************",
                                "",
                                "TASK [nova-cell : Reload nova cell services to remove RPC version cap] *********",
                                "Friday 13 March 2026  01:15:03 +0000 (0:00:00.514)       0:09:04.556 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova API services] *****************************************",
                                "",
                                "TASK [nova : Reload nova API services to remove RPC version pin] ***************",
                                "Friday 13 March 2026  01:15:04 +0000 (0:00:01.253)       0:09:05.810 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Run Nova API online data migrations] *************************************",
                                "",
                                "TASK [nova : Run Nova API online database migrations] **************************",
                                "Friday 13 March 2026  01:15:05 +0000 (0:00:00.723)       0:09:06.534 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Run Nova cell online data migrations] ************************************",
                                "",
                                "TASK [nova-cell : Run Nova cell online database migrations] ********************",
                                "Friday 13 March 2026  01:15:05 +0000 (0:00:00.807)       0:09:07.341 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=59  \u001b[0m \u001b[0;33mchanged=39  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=49  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=56  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=32  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=56  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=29  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=26  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=28  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:15:06 +0000 (0:00:00.454)       0:09:07.795 ********** ",
                                "=============================================================================== ",
                                "nova-cell : Restart nova-libvirt container ----------------------------- 32.43s",
                                "nova : Running Nova API bootstrap container ---------------------------- 29.00s",
                                "nova-cell : Restart nova-compute container ----------------------------- 22.84s",
                                "nova-cell : Restart nova-ssh container --------------------------------- 21.86s",
                                "nova-cell : Running Nova cell bootstrap container ---------------------- 21.11s",
                                "nova-cell : Waiting for nova-compute services to register themselves --- 20.05s",
                                "nova-cell : Get new Libvirt version ------------------------------------ 18.87s",
                                "nova : Restart nova-scheduler container -------------------------------- 18.80s",
                                "nova : Running Nova API bootstrap container ---------------------------- 17.83s",
                                "nova : Create cell0 mappings ------------------------------------------- 14.83s",
                                "nova-cell : Get a list of existing cells ------------------------------- 13.13s",
                                "nova-cell : Get a list of existing cells ------------------------------- 12.98s",
                                "nova-cell : Discover nova hosts ---------------------------------------- 12.96s",
                                "nova-cell : Create cell ------------------------------------------------ 11.93s",
                                "nova-cell : Restart nova-novncproxy container -------------------------- 11.48s",
                                "nova-cell : Get a list of existing cells ------------------------------- 11.15s",
                                "service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 9.88s",
                                "nova-cell : Fail if nova-compute service failed to register ------------- 9.72s",
                                "nova-cell : Get container facts ----------------------------------------- 8.54s",
                                "service-ks-register : nova | Granting/revoking user roles --------------- 8.22s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:10:11 +0000 (0:00:00.192)       0:00:00.192 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:10:12 +0000 (0:00:00.258)       0:00:00.451 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m",
                                "",
                                "PLAY [Apply role octavia] ******************************************************",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:10:12 +0000 (0:00:00.352)       0:00:00.804 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating/deleting services] **************",
                                "Friday 13 March 2026  01:10:13 +0000 (0:00:00.444)       0:00:01.248 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating/deleting endpoints] *************",
                                "Friday 13 March 2026  01:10:16 +0000 (0:00:03.572)       0:00:04.820 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating projects] ***********************",
                                "Friday 13 March 2026  01:10:22 +0000 (0:00:06.287)       0:00:11.108 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating users] **************************",
                                "Friday 13 March 2026  01:10:25 +0000 (0:00:02.773)       0:00:13.882 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating roles] **************************",
                                "Friday 13 March 2026  01:10:32 +0000 (0:00:06.910)       0:00:20.793 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Granting/revoking user roles] ************",
                                "Friday 13 March 2026  01:10:35 +0000 (0:00:03.083)       0:00:23.876 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "",
                                "TASK [octavia : Adding octavia related roles] **********************************",
                                "Friday 13 March 2026  01:10:41 +0000 (0:00:06.223)       0:00:30.099 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:10:56 +0000 (0:00:14.932)       0:00:45.032 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create amphora flavor] *****************************************",
                                "Friday 13 March 2026  01:10:57 +0000 (0:00:00.538)       0:00:45.571 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create nova keypair for amphora] *******************************",
                                "Friday 13 March 2026  01:11:02 +0000 (0:00:05.287)       0:00:50.858 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Friday 13 March 2026  01:11:06 +0000 (0:00:03.972)       0:00:54.831 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create security groups for octavia] ****************************",
                                "Friday 13 March 2026  01:11:09 +0000 (0:00:02.887)       0:00:57.718 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Add rules for security groups] *********************************",
                                "Friday 13 March 2026  01:11:18 +0000 (0:00:08.871)       0:01:06.589 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management network] ************************",
                                "Friday 13 March 2026  01:11:32 +0000 (0:00:13.854)       0:01:20.444 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management subnet] *************************",
                                "Friday 13 March 2026  01:11:36 +0000 (0:00:04.292)       0:01:24.737 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management router for IPv6] ****************",
                                "Friday 13 March 2026  01:11:41 +0000 (0:00:04.631)       0:01:29.368 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update loadbalancer management subnet] *************************",
                                "Friday 13 March 2026  01:11:41 +0000 (0:00:00.270)       0:01:29.639 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:11:45 +0000 (0:00:04.190)       0:01:33.829 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-1, testbed-node-0, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create ports for Octavia health-manager nodes] *****************",
                                "Friday 13 March 2026  01:11:46 +0000 (0:00:01.085)       0:01:34.914 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Update Octavia health manager port host_id] ********************",
                                "Friday 13 March 2026  01:11:53 +0000 (0:00:06.380)       0:01:41.295 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Add Octavia port to openvswitch br-int] ************************",
                                "Friday 13 March 2026  01:11:57 +0000 (0:00:04.125)       0:01:45.420 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Install isc-dhcp-client package] *******************************",
                                "Friday 13 March 2026  01:11:58 +0000 (0:00:00.831)       0:01:46.251 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia dhclient conf] **********************************",
                                "Friday 13 March 2026  01:12:00 +0000 (0:00:02.366)       0:01:48.618 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia-interface service] ******************************",
                                "Friday 13 March 2026  01:12:01 +0000 (0:00:01.432)       0:01:50.050 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Restart octavia-interface.service if required] *****************",
                                "Friday 13 March 2026  01:12:03 +0000 (0:00:01.289)       0:01:51.340 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Enable and start octavia-interface.service] ********************",
                                "Friday 13 March 2026  01:12:04 +0000 (0:00:01.840)       0:01:53.180 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Wait for interface ohm0 ip appear] *****************************",
                                "Friday 13 March 2026  01:12:06 +0000 (0:00:01.930)       0:01:55.111 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Gather facts] **************************************************",
                                "Friday 13 March 2026  01:12:07 +0000 (0:00:00.625)       0:01:55.736 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:12:10 +0000 (0:00:02.937)       0:01:58.674 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Get amphora flavor info] ***************************************",
                                "Friday 13 March 2026  01:12:11 +0000 (0:00:00.853)       0:01:59.528 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Friday 13 March 2026  01:12:15 +0000 (0:00:04.249)       0:02:03.778 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get security groups for octavia] *******************************",
                                "Friday 13 March 2026  01:12:18 +0000 (0:00:03.163)       0:02:06.941 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Get loadbalancer management network] ***************************",
                                "Friday 13 March 2026  01:12:25 +0000 (0:00:06.562)       0:02:13.504 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia resources facts] ***********************************",
                                "Friday 13 March 2026  01:12:28 +0000 (0:00:03.531)       0:02:17.036 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Ensuring config directories exist] *****************************",
                                "Friday 13 March 2026  01:12:29 +0000 (0:00:00.309)       0:02:17.346 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Check if policies shall be overwritten] ************************",
                                "Friday 13 March 2026  01:12:31 +0000 (0:00:02.367)       0:02:19.713 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia policy file] ***************************************",
                                "Friday 13 March 2026  01:12:31 +0000 (0:00:00.122)       0:02:19.836 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over existing policy file] *****************************",
                                "Friday 13 March 2026  01:12:32 +0000 (0:00:00.454)       0:02:20.291 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:12:32 +0000 (0:00:00.665)       0:02:20.957 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over extra CA certificates] ********",
                                "Friday 13 March 2026  01:12:33 +0000 (0:00:00.544)       0:02:21.501 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***",
                                "Friday 13 March 2026  01:12:38 +0000 (0:00:05.122)       0:02:26.624 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****",
                                "Friday 13 March 2026  01:12:39 +0000 (0:00:00.936)       0:02:27.560 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over config.json files for services] *******************",
                                "Friday 13 March 2026  01:12:40 +0000 (0:00:00.876)       0:02:28.436 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia-wsgi.conf] ********************************",
                                "Friday 13 March 2026  01:12:45 +0000 (0:00:04.830)       0:02:33.267 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia.conf] *************************************",
                                "Friday 13 March 2026  01:12:46 +0000 (0:00:01.779)       0:02:35.046 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over Octavia SSH key] **********************************",
                                "Friday 13 March 2026  01:13:03 +0000 (0:00:16.588)       0:02:51.635 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-worker] ******************",
                                "Friday 13 March 2026  01:13:04 +0000 (0:00:01.303)       0:02:52.938 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-housekeeping] ************",
                                "Friday 13 March 2026  01:13:09 +0000 (0:00:04.480)       0:02:57.419 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-health-manager] **********",
                                "Friday 13 March 2026  01:13:14 +0000 (0:00:05.091)       0:03:02.511 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [service-check-containers : octavia | Check containers] *******************",
                                "Friday 13 March 2026  01:13:18 +0000 (0:00:04.624)       0:03:07.136 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-check-containers : octavia | Notify handlers to restart containers] ***",
                                "Friday 13 March 2026  01:13:22 +0000 (0:00:03.663)       0:03:10.800 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;33m    \"msg\": \"Notifying handlers\"\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "TASK [service-check-containers : Include tasks] ********************************",
                                "Friday 13 March 2026  01:13:22 +0000 (0:00:00.346)       0:03:11.146 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2025.1', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2025.1', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2025.1', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2025.1', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2025.1', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/dev/shm:/dev/shm', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:13:24 +0000 (0:00:01.290)       0:03:12.436 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database] *************************************",
                                "Friday 13 March 2026  01:13:24 +0000 (0:00:00.301)       0:03:12.738 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database] *************************",
                                "Friday 13 March 2026  01:13:26 +0000 (0:00:02.176)       0:03:14.914 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database user and setting permissions] ********",
                                "Friday 13 March 2026  01:13:28 +0000 (0:00:02.108)       0:03:17.022 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database user and setting permissions] ***",
                                "Friday 13 March 2026  01:13:30 +0000 (0:00:02.056)       0:03:19.079 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Running Octavia bootstrap container] ***************************",
                                "Friday 13 March 2026  01:13:32 +0000 (0:00:01.841)       0:03:20.920 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:13:54 +0000 (0:00:21.553)       0:03:42.474 ********** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:13:54 +0000 (0:00:00.069)       0:03:42.543 ********** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Friday 13 March 2026  01:13:54 +0000 (0:00:00.068)       0:03:42.611 ********** ",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-api container] **********************",
                                "Friday 13 March 2026  01:13:54 +0000 (0:00:00.265)       0:03:42.877 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************",
                                "Friday 13 March 2026  01:14:08 +0000 (0:00:14.044)       0:03:56.922 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********",
                                "Friday 13 March 2026  01:14:19 +0000 (0:00:10.911)       0:04:07.834 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************",
                                "Friday 13 March 2026  01:14:31 +0000 (0:00:11.756)       0:04:19.591 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-worker container] *******************",
                                "Friday 13 March 2026  01:14:42 +0000 (0:00:10.783)       0:04:30.374 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=58  \u001b[0m \u001b[0;33mchanged=39  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:14:47 +0000 (0:00:05.088)       0:04:35.462 ********** ",
                                "=============================================================================== ",
                                "octavia : Running Octavia bootstrap container -------------------------- 21.55s",
                                "octavia : Copying over octavia.conf ------------------------------------ 16.59s",
                                "octavia : Adding octavia related roles --------------------------------- 14.93s",
                                "octavia : Restart octavia-api container -------------------------------- 14.04s",
                                "octavia : Add rules for security groups -------------------------------- 13.85s",
                                "octavia : Restart octavia-health-manager container --------------------- 11.76s",
                                "octavia : Restart octavia-driver-agent container ----------------------- 10.91s",
                                "octavia : Restart octavia-housekeeping container ----------------------- 10.78s",
                                "octavia : Create security groups for octavia ---------------------------- 8.87s",
                                "service-ks-register : octavia | Creating users -------------------------- 6.91s",
                                "octavia : Get security groups for octavia ------------------------------- 6.56s",
                                "octavia : Create ports for Octavia health-manager nodes ----------------- 6.38s",
                                "service-ks-register : octavia | Creating/deleting endpoints ------------- 6.29s",
                                "service-ks-register : octavia | Granting/revoking user roles ------------ 6.22s",
                                "octavia : Create amphora flavor ----------------------------------------- 5.29s",
                                "service-cert-copy : octavia | Copying over extra CA certificates -------- 5.12s",
                                "octavia : Copying certificate files for octavia-housekeeping ------------ 5.09s",
                                "octavia : Restart octavia-worker container ------------------------------ 5.09s",
                                "octavia : Copying over config.json files for services ------------------- 4.83s",
                                "octavia : Create loadbalancer management subnet ------------------------- 4.63s",
                                "\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b050cd2-9f80-4de7-a218-c54db7fdb4d0 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- END -- Fri Mar 13 01:16:26 UTC 2026"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001c-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:16:26.800936Z",
                            "start": "2026-03-13T00:42:32.164810Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001c",
                        "name": "Deploy in a nutshell"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "/opt/configuration/scripts/bootstrap.sh"
                            ],
                            "delta": "0:02:50.419065",
                            "end": "2026-03-13 01:19:17.659088",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 /opt/configuration/scripts/bootstrap.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001d-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 01:16:27.240023",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# BOOTSTRAP\n\n+ set -e\n+ echo\n+ echo '# BOOTSTRAP'\n+ echo\n+ sh -c /opt/configuration/scripts/bootstrap-services.sh\n+ set -e\n+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh\n\u001b[32m2026-03-13 01:16:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 73f246c2-f544-47b1-8f3d-5cd1f22d8481 (flavor-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-13 01:16:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m\n\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m\n\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m\n\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m\n\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m\n\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m\n\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m\n\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-100s created\u001b[0m\n\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m\n\u001b[32m2026-03-13 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m\n\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m\n\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9ecd8acd-8c50-4bbd-b299-1f0ca70a8292 (bootstrap-basic) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9ecd8acd-8c50-4bbd-b299-1f0ca70a8292 (bootstrap-basic) has been started and output is visible here.\u001b[0m\n\nPLAY [Bootstrap basic OpenStack services] **************************************\n\nTASK [Gathering Facts] *********************************************************\nFriday 13 March 2026  01:16:55 +0000 (0:00:00.062)       0:00:00.062 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Get volume type LUKS] ****************************************************\nFriday 13 March 2026  01:16:57 +0000 (0:00:01.802)       0:00:01.865 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Create volume type LUKS] *************************************************\nFriday 13 March 2026  01:17:05 +0000 (0:00:07.888)       0:00:09.753 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public network] ***************************************************\nFriday 13 March 2026  01:17:12 +0000 (0:00:06.854)       0:00:16.608 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Set public network to default] *******************************************\nFriday 13 March 2026  01:17:17 +0000 (0:00:05.075)       0:00:21.684 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public subnet] ****************************************************\nFriday 13 March 2026  01:17:23 +0000 (0:00:06.180)       0:00:27.864 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create default IPv4 subnet pool] *****************************************\nFriday 13 March 2026  01:17:27 +0000 (0:00:03.999)       0:00:31.864 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create manager role] *****************************************************\nFriday 13 March 2026  01:17:31 +0000 (0:00:03.662)       0:00:35.527 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:17:34 +0000 (0:00:03.401)       0:00:38.929 ********** \n=============================================================================== \nGet volume type LUKS ---------------------------------------------------- 7.89s\nCreate volume type LUKS ------------------------------------------------- 6.85s\nSet public network to default ------------------------------------------- 6.18s\nCreate public network --------------------------------------------------- 5.08s\nCreate public subnet ---------------------------------------------------- 4.00s\nCreate default IPv4 subnet pool ----------------------------------------- 3.66s\nCreate manager role ----------------------------------------------------- 3.40s\nGathering Facts --------------------------------------------------------- 1.80s\n\u001b[32m2026-03-13 01:17:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06e7cacc-7f38-4b9b-8c9b-bf1b42835f9c (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m\n\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-13 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-13 01:17:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m\n\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-13 01:17:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-13 01:18:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m\n\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m\n\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m\n\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m\n+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m\n\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m\n\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2.CHECKSUM\u001b[0m\n\u001b[32m2026-03-13 01:18:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: b71c87d72a1bb2a551beefb0dd91604facfc9dbadb0f72aab9f33bc3720b0c7c\u001b[0m\n\u001b[32m2026-03-13 01:18:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7e23bc33-85a9-4993-8079-a4c4bc59b33c (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m\n\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2: 200\u001b[0m\n\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m\n\u001b[32m2026-03-13 01:18:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-03-13 01:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:18:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:18:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:19:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-12' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m\n\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m\n\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m\n\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m\n\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-12\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m\n\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m",
                            "stdout_lines": [
                                "",
                                "# BOOTSTRAP",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# BOOTSTRAP'",
                                "+ echo",
                                "+ sh -c /opt/configuration/scripts/bootstrap-services.sh",
                                "+ set -e",
                                "+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh",
                                "\u001b[32m2026-03-13 01:16:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 73f246c2-f544-47b1-8f3d-5cd1f22d8481 (flavor-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-100s created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9ecd8acd-8c50-4bbd-b299-1f0ca70a8292 (bootstrap-basic) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 01:16:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 9ecd8acd-8c50-4bbd-b299-1f0ca70a8292 (bootstrap-basic) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Bootstrap basic OpenStack services] **************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Friday 13 March 2026  01:16:55 +0000 (0:00:00.062)       0:00:00.062 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Get volume type LUKS] ****************************************************",
                                "Friday 13 March 2026  01:16:57 +0000 (0:00:01.802)       0:00:01.865 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Create volume type LUKS] *************************************************",
                                "Friday 13 March 2026  01:17:05 +0000 (0:00:07.888)       0:00:09.753 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public network] ***************************************************",
                                "Friday 13 March 2026  01:17:12 +0000 (0:00:06.854)       0:00:16.608 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Set public network to default] *******************************************",
                                "Friday 13 March 2026  01:17:17 +0000 (0:00:05.075)       0:00:21.684 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public subnet] ****************************************************",
                                "Friday 13 March 2026  01:17:23 +0000 (0:00:06.180)       0:00:27.864 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create default IPv4 subnet pool] *****************************************",
                                "Friday 13 March 2026  01:17:27 +0000 (0:00:03.999)       0:00:31.864 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create manager role] *****************************************************",
                                "Friday 13 March 2026  01:17:31 +0000 (0:00:03.662)       0:00:35.527 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:17:34 +0000 (0:00:03.401)       0:00:38.929 ********** ",
                                "=============================================================================== ",
                                "Get volume type LUKS ---------------------------------------------------- 7.89s",
                                "Create volume type LUKS ------------------------------------------------- 6.85s",
                                "Set public network to default ------------------------------------------- 6.18s",
                                "Create public network --------------------------------------------------- 5.08s",
                                "Create public subnet ---------------------------------------------------- 4.00s",
                                "Create default IPv4 subnet pool ----------------------------------------- 3.66s",
                                "Create manager role ----------------------------------------------------- 3.40s",
                                "Gathering Facts --------------------------------------------------------- 1.80s",
                                "\u001b[32m2026-03-13 01:17:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 06e7cacc-7f38-4b9b-8c9b-bf1b42835f9c (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-13 01:17:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2.CHECKSUM\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: b71c87d72a1bb2a551beefb0dd91604facfc9dbadb0f72aab9f33bc3720b0c7c\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7e23bc33-85a9-4993-8079-a4c4bc59b33c (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2: 200\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:18:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-03-12' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2025.1.20260312.qcow2\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-03-12\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-03-12'\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:17\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001d-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:19:18.103365Z",
                            "start": "2026-03-13T01:16:26.997086Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001d",
                        "name": "Bootstrap services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "/opt/configuration/scripts/check.sh"
                            ],
                            "delta": "0:17:28.703905",
                            "end": "2026-03-13 01:36:47.149951",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 /opt/configuration/scripts/check.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001e-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 01:19:18.446046",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# CHECK\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# CHECK'\n+ echo\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-manager\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-manager'\n+ echo\n+ osism container testbed-manager ps\n\u001b[32m2026-03-13 01:19:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m\nCONTAINER ID   IMAGE                                                           COMMAND                  CREATED          STATUS                    PORTS                                                                             NAMES\n6c9c3f91c326   registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_blackbox_exporter\n808be0674e46   registry.osism.tech/kolla/prometheus-alertmanager:2025.1        \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_alertmanager\ndf0c79486a88   registry.osism.tech/kolla/prometheus-cadvisor:2025.1            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_cadvisor\n0bf2b9aedfe6   registry.osism.tech/kolla/prometheus-node-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                                                                                               prometheus_node_exporter\n72fc99802dae   registry.osism.tech/kolla/prometheus-server:2025.1              \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                                                                                               prometheus_server\n0487d415e717   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   18 minutes ago   Up 17 minutes                                                                                               cephclient\nd6aa562df796   registry.osism.tech/kolla/cron:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               cron\n0111d7689ed8   registry.osism.tech/kolla/kolla-toolbox:2025.1                  \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               kolla_toolbox\n0b2f1a7a6a02   registry.osism.tech/kolla/fluentd:2025.1                        \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                                                                                               fluentd\nacfaccc59dd1   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   30 minutes ago   Up 29 minutes (healthy)   80/tcp                                                                            phpmyadmin\ncc26c373effc   registry.osism.tech/osism/openstackclient:2025.1                \"/usr/bin/dumb-init \u2026\"   31 minutes ago   Up 30 minutes                                                                                               openstackclient\na0520e08f520   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   31 minutes ago   Up 30 minutes (healthy)   8080/tcp                                                                          homer\n10e81a04583b   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   54 minutes ago   Up 53 minutes (healthy)   192.168.16.5:3128->3128/tcp                                                       squid\nfaeb53a8dc03   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     manager-inventory_reconciler-1\nec56de84559b   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     osism-kubernetes\n913c2ab24057   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     osism-ansible\n8a07529d0df1   registry.osism.tech/osism/kolla-ansible:2025.1                  \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     kolla-ansible\n06851db83dd9   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     ceph-ansible\n3d43f54011a4   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   58 minutes ago   Up 38 minutes (healthy)   8000/tcp                                                                          manager-ara-server-1\n093c1ecb28f9   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes (healthy)   6379/tcp                                                                          manager-redis-1\n1903eeaddbd6   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-flower-1\n9c3b1270a042   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes (healthy)   3306/tcp                                                                          manager-mariadb-1\n1fca8a62a0b5   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-openstack-1\n2c059a1d2a71   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-beat-1\nf924682c1247   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-listener-1\nd58e9b803ce9   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes             192.168.16.5:3000->3000/tcp                                                       osism-frontend\n70ff0fbdeaa8   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1\ncaf640858b6b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     osismclient\nafba45b5e55b   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   59 minutes ago   Up 59 minutes (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik\n\n## Images @ testbed-manager\n\n+ echo\n+ echo '## Images @ testbed-manager'\n+ echo\n+ osism container testbed-manager images\nREPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE\nregistry.osism.tech/osism/osism-ansible                  latest           3c82227097a8   About an hour ago   613MB\nregistry.osism.tech/osism/kolla-ansible                  2025.1           385f40f5bb87   About an hour ago   610MB\nregistry.osism.tech/osism/osism                          latest           3376395f1aa5   About an hour ago   406MB\nregistry.osism.tech/osism/ceph-ansible                   reef             c0b375cdf34b   About an hour ago   560MB\nregistry.osism.tech/osism/osism-kubernetes               latest           bb68cd12c222   About an hour ago   1.22GB\nregistry.osism.tech/osism/osism-frontend                 latest           2a698a7d7ce5   About an hour ago   232MB\nregistry.osism.tech/osism/inventory-reconciler           latest           0dc880827435   About an hour ago   335MB\nregistry.osism.tech/osism/openstackclient                2025.1           26b681b56b01   22 hours ago        212MB\nregistry.osism.tech/osism/cephclient                     reef             19951ee12217   22 hours ago        453MB\nregistry.osism.tech/kolla/kolla-toolbox                  2025.1           a7d5f8f5723c   23 hours ago        677MB\nregistry.osism.tech/kolla/fluentd                        2025.1           57639eec9828   23 hours ago        585MB\nregistry.osism.tech/kolla/cron                           2025.1           69f3c33abd1e   23 hours ago        271MB\nregistry.osism.tech/kolla/prometheus-server              2025.1           cdeb71ca24f7   23 hours ago        855MB\nregistry.osism.tech/kolla/prometheus-blackbox-exporter   2025.1           2a08c0f63ea9   23 hours ago        314MB\nregistry.osism.tech/kolla/prometheus-cadvisor            2025.1           98f129deee0e   23 hours ago        363MB\nregistry.osism.tech/kolla/prometheus-alertmanager        2025.1           6d47b066a3eb   23 hours ago        409MB\nregistry.osism.tech/kolla/prometheus-node-exporter       2025.1           ce49e82a896f   23 hours ago        311MB\nregistry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   6 weeks ago         41.4MB\nregistry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   3 months ago        11.5MB\nregistry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   3 months ago        334MB\nphpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   5 months ago        742MB\nregistry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB\nregistry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   7 months ago        226MB\nregistry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   21 months ago       146MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-0\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\nd7a63897fdbd   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\na7158a6cc794   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n1208e6186247   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n407e2d2c5e67   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\n071c7d07fc86   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\n3fa07550e17b   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   6 minutes ago    Up 6 minutes (healthy)              nova_novncproxy\n02a2b2aa767b   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\nc3e9eebc5b69   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata\ncc4cf05d3c43   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api\nedfb39ff2955   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n49ea21131d32   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\n65c8f3fa6eb4   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       grafana\n5d7214f53bdd   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\ncb4bc32826a9   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\n70870ae1ee11   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\ndd0afc8356d0   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\nf5500f3411e3   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter\n4f542e28f74c   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor\n944b8b59133b   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\neab91fc6e472   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\nec67d88396ea   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter\n871651d1bc43   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor\n6c08ddc7ac34   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api\n234d84c97c56   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server\n4705fcd70466   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api\n4018c9f2e667   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker\n37ceb648b4ec   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns\n5bd0f24ccb91   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer\n3a8b6f299bfe   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central\n482b42f701c6   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api\n8a80fea66d1f   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\n4c3cf114c1f4   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker\neb1031c2c652   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener\ndc2ebb5b50c1   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api\nad46f532a250   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-0\na395f068dba4   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\nb0ef065ac852   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\n8190cbadd291   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\n4e6e6a06584a   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             horizon\n15822dfe46e3   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb\n2fcf94ade88a   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards\n445c7143fd80   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch\n64a2621e7382   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-0\nc6bf31483a83   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\n66bad52d2efe   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\n8aef63697d89   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\n0bf63e175326   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd\n055dec7c911d   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1\ne7a60857f864   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_sb_db\n27298373da9d   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_nb_db\ne9e57f170ae9   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\n8f8762892bf0   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-0\n709432bb4858   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             rabbitmq\n0a94b45018f0   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n20a365466b20   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_db\nd1647407d1e9   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n8b18a340cf76   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\n56aacb432599   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\ncf316d177793   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\n6538d643b237   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox\ne47483949a18   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-0\n\n+ echo\n+ echo '## Images @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB\nregistry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB\nregistry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB\nregistry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB\nregistry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB\nregistry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB\nregistry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB\nregistry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB\nregistry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB\nregistry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB\nregistry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB\nregistry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB\nregistry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB\nregistry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB\nregistry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB\nregistry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB\nregistry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB\nregistry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB\nregistry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB\nregistry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB\nregistry.osism.tech/kolla/ceilometer-notification             2025.1    ece9531359a7   23 hours ago   990MB\nregistry.osism.tech/kolla/ceilometer-central                  2025.1    85a92ad61831   23 hours ago   991MB\nregistry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB\nregistry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB\nregistry.osism.tech/kolla/aodh-api                            2025.1    cbdf1972f1eb   23 hours ago   988MB\nregistry.osism.tech/kolla/aodh-evaluator                      2025.1    a958cce06a38   23 hours ago   989MB\nregistry.osism.tech/kolla/aodh-notifier                       2025.1    0c3a6cf5d0cf   23 hours ago   989MB\nregistry.osism.tech/kolla/aodh-listener                       2025.1    e9a41838f69b   23 hours ago   989MB\nregistry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB\nregistry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB\nregistry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB\nregistry.osism.tech/kolla/skyline-console                     2025.1    ebd12ce3748e   23 hours ago   1.06GB\nregistry.osism.tech/kolla/skyline-apiserver                   2025.1    d463dff27fc9   23 hours ago   1.01GB\nregistry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB\nregistry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB\nregistry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB\nregistry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB\nregistry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-1\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n3a0122e25fb1   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\n5dfe2326c2b0   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n024ef5c3f0d4   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n013ec66becca   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\nf61f87918eb2   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\n9194f09ba2d8   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   6 minutes ago    Up 6 minutes (healthy)              nova_novncproxy\n4a481e4b3739   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\n6b60b417b766   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata\n940a90b42bac   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api\n4ecfee1e6c47   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n8acfc72e80bf   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana\n0dd4b999a79c   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\na406a59c5bd8   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\n3255c40a34e1   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\n7bab09b9da97   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\n4e16feaea168   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\nfd9991dd1999   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter\n517d8d616064   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor\nf072ca386a63   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\n3d599f288bee   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\n0ffd8d0b7ef9   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter\n241b36fba680   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor\nb1736d9cea81   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api\ne38e68095a28   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server\n9451dc358496   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api\n3b79dea4bbda   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker\n1182c997c693   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns\n17849d7f8f56   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer\nf95bb464161d   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central\n7fcf8a972a49   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api\na56c14220e7e   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\n25548567dfff   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker\n71790b96b08e   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_keystone_listener\n1c9abb6281d6   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api\nb28d19404e6b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-1\n46caf03bfa65   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\n79fea1e636a9   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\n5a69b4644354   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon\n5df73378b3f8   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\n0859ecf8b321   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards\n6f5c4963924d   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb\n7d7816f04e61   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 21 minutes (healthy)             opensearch\nc922c44eca4b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-1\n7b9b7c77f504   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\n1327ecce99d2   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\n0fe32ee24f09   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\nf525617ce17e   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd\nda9b94edc26c   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq\nbe63fc264077   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1\ndfbe085cd6fa   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db\n00e07a8211f2   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_nb_db\nb99614e57c6e   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\ncb0a6fe177a7   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-1\n4c6ae9cc1d7d   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n94f1b5442875   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_db\nb3670fa142db   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n6e430df82abd   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\nda2d00ec0ae4   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\nb2ce6bac3ad5   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\n8a1e8d661817   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox\nee7b7af32b6c   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-1\n\n+ echo\n+ echo '## Images @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB\nregistry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB\nregistry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB\nregistry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB\nregistry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB\nregistry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB\nregistry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB\nregistry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB\nregistry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB\nregistry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB\nregistry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB\nregistry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB\nregistry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB\nregistry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB\nregistry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB\nregistry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB\nregistry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB\nregistry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB\nregistry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB\nregistry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB\nregistry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB\nregistry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB\nregistry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB\nregistry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB\nregistry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB\nregistry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB\nregistry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB\nregistry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB\nregistry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB\nregistry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n\n## Containers @ testbed-node-2\n\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\ne6817ae3c070   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\ncd5f49b10bcd   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_housekeeping\n344287924d1f   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n8462bbab99fd   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\nf7ed50e200d2   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\n372277f6f27b   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy\na895ff617c8b   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor\ndd49ed6635c6   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata\n9097404d4703   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api\n0efb91e22778   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler\n222776286f9d   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   11 minutes ago   Up 10 minutes                       grafana\n872c63553423   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\nb532d81835ac   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\n6fd57c357f33   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler\ne40ebd44e115   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api\n585e27543ffa   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\n224c4450d4d9   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter\n7185f150b760   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor\ndf0ed0ec176a   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\nac935971b7fe   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\n6569728c8720   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter\n7610a3e2d08b   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor\nd229ebe45517   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api\nff3788c1b8b8   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server\n2722ff7750bc   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api\nceea4d26aa1f   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker\n5a28a4352cb4   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns\nb414321f3049   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer\nf794b9f6ff79   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central\n609bba6b2a0f   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api\n364c282522f8   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9\nebe73c91b961   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_worker\n571859c31679   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_keystone_listener\n15af3a2eeaa5   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-2\n62b6bb95b695   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api\nf2b97b58356e   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone\n98954b0e8036   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet\nf21f0479608c   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon\n53b872fd5cdf   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh\n0ece719a6c89   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb\n5e491dcd59f9   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards\nb055fca7034b   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch\n1479c49abc78   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-2\n7a772282ae0b   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived\nabdf42e8302a   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql\n0a869a760f28   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy\nbe42f02e5b3c   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq\ndeb57168100c   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd\nd4052e508b85   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1\n0028e2095a55   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db\nde0d08a3fa58   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_nb_db\n40f00350cb17   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller\n73131559e39b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   28 minutes ago   Up 28 minutes                       ceph-mon-testbed-node-2\nffcfd9858ab9   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd\n3801462a5332   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db\n2785382df2cf   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel\n686809c29d19   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis\nc1265b35218f   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached\n377523689e65   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron\n84642d628a7d   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox\n7398de513872   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd\n\n## Images @ testbed-node-2\n\n+ echo\n+ echo '## Images @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB\nregistry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB\nregistry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB\nregistry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB\nregistry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB\nregistry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB\nregistry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB\nregistry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB\nregistry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB\nregistry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB\nregistry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB\nregistry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB\nregistry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB\nregistry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB\nregistry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB\nregistry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB\nregistry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB\nregistry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB\nregistry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB\nregistry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB\nregistry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB\nregistry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB\nregistry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB\nregistry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB\nregistry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB\nregistry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB\nregistry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB\nregistry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB\nregistry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB\nregistry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB\nregistry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB\nregistry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB\nregistry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB\nregistry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB\nregistry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB\nregistry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB\nregistry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB\n+ sh -c /opt/configuration/scripts/check-services.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# Ceph status\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# Ceph status'\n+ echo\n+ ceph -s\n  cluster:\n    id:     11111111-1111-1111-1111-111111111111\n    health: HEALTH_OK\n \n  services:\n    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 27m)\n    mgr: testbed-node-0(active, since 16m), standbys: testbed-node-2, testbed-node-1\n    mds: 1/1 daemons up, 2 standby\n    osd: 6 osds: 6 up (since 24m), 6 in (since 24m)\n    rgw: 3 daemons active (3 hosts, 1 zones)\n \n  data:\n    volumes: 1/1 healthy\n    pools:   14 pools, 401 pgs\n    objects: 556 objects, 2.2 GiB\n    usage:   7.1 GiB used, 113 GiB / 120 GiB avail\n    pgs:     401 active+clean\n \n\n# Ceph versions\n\n+ echo\n+ echo '# Ceph versions'\n+ echo\n+ ceph versions\n{\n    \"mon\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"mgr\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"osd\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6\n    },\n    \"mds\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"rgw\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"overall\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18\n    }\n}\n\n# Ceph OSD tree\n\n+ echo\n+ echo '# Ceph OSD tree'\n+ echo\n+ ceph osd df tree\nID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA      OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              \n-1         0.11691         -  120 GiB  7.1 GiB   6.7 GiB    6 KiB  430 MiB  113 GiB  5.92  1.00    -          root default           \n-3         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-3\n 1    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.2 GiB    1 KiB   70 MiB   19 GiB  6.45  1.09  204      up          osd.1          \n 4    hdd  0.01949   1.00000   20 GiB  1.1 GiB   1.0 GiB    1 KiB   74 MiB   19 GiB  5.38  0.91  186      up          osd.4          \n-7         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4\n 0    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.2 GiB    1 KiB   74 MiB   19 GiB  6.59  1.11  175      up          osd.0          \n 3    hdd  0.01949   1.00000   20 GiB  1.0 GiB  1003 MiB    1 KiB   70 MiB   19 GiB  5.24  0.89  213      up          osd.3          \n-5         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-5\n 2    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.3 GiB    1 KiB   70 MiB   19 GiB  6.69  1.13  206      up          osd.2          \n 5    hdd  0.01949   1.00000   20 GiB  1.0 GiB   979 MiB    1 KiB   74 MiB   19 GiB  5.14  0.87  186      up          osd.5          \n                       TOTAL  120 GiB  7.1 GiB   6.7 GiB  9.3 KiB  430 MiB  113 GiB  5.92                                            \nMIN/MAX VAR: 0.87/1.13  STDDEV: 0.67\n\n# Ceph monitor status\n\n+ echo\n+ echo '# Ceph monitor status'\n+ echo\n+ ceph mon stat\ne1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2\n\n# Ceph quorum status\n\n+ echo\n+ echo '# Ceph quorum status'\n+ echo\n+ ceph quorum_status\n+ jq\n{\n  \"election_epoch\": 8,\n  \"quorum\": [\n    0,\n    1,\n    2\n  ],\n  \"quorum_names\": [\n    \"testbed-node-0\",\n    \"testbed-node-1\",\n    \"testbed-node-2\"\n  ],\n  \"quorum_leader_name\": \"testbed-node-0\",\n  \"quorum_age\": 1680,\n  \"features\": {\n    \"quorum_con\": \"4540138322906710015\",\n    \"quorum_mon\": [\n      \"kraken\",\n      \"luminous\",\n      \"mimic\",\n      \"osdmap-prune\",\n      \"nautilus\",\n      \"octopus\",\n      \"pacific\",\n      \"elector-pinging\",\n      \"quincy\",\n      \"reef\"\n    ]\n  },\n  \"monmap\": {\n    \"epoch\": 1,\n    \"fsid\": \"11111111-1111-1111-1111-111111111111\",\n    \"modified\": \"2026-03-13T00:51:25.351727Z\",\n    \"created\": \"2026-03-13T00:51:25.351727Z\",\n    \"min_mon_release\": 18,\n    \"min_mon_release_name\": \"reef\",\n    \"election_strategy\": 1,\n    \"disallowed_leaders: \": \"\",\n    \"stretch_mode\": false,\n    \"tiebreaker_mon\": \"\",\n    \"removed_ranks: \": \"\",\n    \"features\": {\n      \"persistent\": [\n        \"kraken\",\n        \"luminous\",\n        \"mimic\",\n        \"osdmap-prune\",\n        \"nautilus\",\n        \"octopus\",\n        \"pacific\",\n        \"elector-pinging\",\n        \"quincy\",\n        \"reef\"\n      ],\n      \"optional\": []\n    },\n    \"mons\": [\n      {\n        \"rank\": 0,\n        \"name\": \"testbed-node-0\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.10:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.10:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.10:6789/0\",\n        \"public_addr\": \"192.168.16.10:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 1,\n        \"name\": \"testbed-node-1\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.11:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.11:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.11:6789/0\",\n        \"public_addr\": \"192.168.16.11:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 2,\n        \"name\": \"testbed-node-2\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.12:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.12:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.12:6789/0\",\n        \"public_addr\": \"192.168.16.12:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      }\n    ]\n  }\n}\n\n# Ceph free space status\n\n+ echo\n+ echo '# Ceph free space status'\n+ echo\n+ ceph df\n--- RAW STORAGE ---\nCLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED\nhdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92\nTOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92\n \n--- POOLS ---\nPOOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL\n.mgr                        1    1  577 KiB        2  1.1 MiB      0     53 GiB\ncephfs_data                 2   32      0 B        0      0 B      0     35 GiB\ncephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB\ndefault.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.control         6   32      0 B        8      0 B      0     35 GiB\ndefault.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB\ndefault.rgw.meta            8   32      0 B        0      0 B      0     35 GiB\n.rgw.root                   9   32  3.9 KiB        8   64 KiB      0     53 GiB\nbackups                    10   32     19 B        2   12 KiB      0     35 GiB\nvolumes                    11   32     19 B        2   12 KiB      0     35 GiB\nimages                     12   32  2.2 GiB      299  6.7 GiB   5.92     35 GiB\nmetrics                    13   32     19 B        2   12 KiB      0     35 GiB\nvms                        14   32     19 B        2   12 KiB      0     35 GiB\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ ! -e /etc/redhat-release ]]\n+ osism apply facts\n\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab5989f-2cab-410b-9a43-0ddd7b56424d (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 6ab5989f-2cab-410b-9a43-0ddd7b56424d (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nFriday 13 March 2026  01:20:00 +0000 (0:00:00.228)       0:00:00.228 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nFriday 13 March 2026  01:20:01 +0000 (0:00:01.326)       0:00:01.554 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nFriday 13 March 2026  01:20:02 +0000 (0:00:01.106)       0:00:02.661 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nFriday 13 March 2026  01:20:07 +0000 (0:00:04.184)       0:00:06.846 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:20:07 +0000 (0:00:00.473)       0:00:07.319 ********** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.18s\nosism.commons.facts : Create custom facts directory --------------------- 1.33s\nosism.commons.facts : Copy fact files ----------------------------------- 1.11s\nGather facts for all hosts ---------------------------------------------- 0.47s\n+ osism validate ceph-mons\n\nPLAY [Ceph validate mons] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nFriday 13 March 2026  01:20:25 +0000 (0:00:00.434)       0:00:00.434 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nFriday 13 March 2026  01:20:25 +0000 (0:00:00.810)       0:00:01.245 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nFriday 13 March 2026  01:20:26 +0000 (0:00:00.916)       0:00:02.161 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nFriday 13 March 2026  01:20:26 +0000 (0:00:00.133)       0:00:02.294 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nFriday 13 March 2026  01:20:27 +0000 (0:00:00.293)       0:00:02.588 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nFriday 13 March 2026  01:20:28 +0000 (0:00:01.219)       0:00:03.808 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nFriday 13 March 2026  01:20:28 +0000 (0:00:00.285)       0:00:04.093 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:20:29 +0000 (0:00:00.471)       0:00:04.565 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mon is not running] ********************\nFriday 13 March 2026  01:20:29 +0000 (0:00:00.301)       0:00:04.867 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mon is running] ************************\nFriday 13 March 2026  01:20:29 +0000 (0:00:00.298)       0:00:05.165 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:20:30 +0000 (0:00:00.442)       0:00:05.607 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:20:30 +0000 (0:00:00.275)       0:00:05.883 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:20:30 +0000 (0:00:00.255)       0:00:06.139 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:30 +0000 (0:00:00.264)       0:00:06.404 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.069)       0:00:06.473 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.067)       0:00:06.540 ********** \n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.088)       0:00:06.629 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.232)       0:00:06.862 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Prepare quorum test vars] ************************************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.234)       0:00:07.097 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get monmap info from one mon container] **********************************\nFriday 13 March 2026  01:20:31 +0000 (0:00:00.122)       0:00:07.220 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set quorum test data] ****************************************************\nFriday 13 March 2026  01:20:33 +0000 (0:00:01.759)       0:00:08.979 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail quorum test if not all monitors are in quorum] **********************\nFriday 13 March 2026  01:20:34 +0000 (0:00:00.476)       0:00:09.456 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass quorum test if all monitors are in quorum] **************************\nFriday 13 March 2026  01:20:34 +0000 (0:00:00.120)       0:00:09.577 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set fsid test vars] ******************************************************\nFriday 13 March 2026  01:20:34 +0000 (0:00:00.335)       0:00:09.913 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail Cluster FSID test if FSID does not match configuration] *************\nFriday 13 March 2026  01:20:34 +0000 (0:00:00.303)       0:00:10.217 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass Cluster FSID test if it matches configuration] **********************\nFriday 13 March 2026  01:20:34 +0000 (0:00:00.110)       0:00:10.327 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare status test vars] ************************************************\nFriday 13 March 2026  01:20:35 +0000 (0:00:00.123)       0:00:10.451 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather status data] ******************************************************\nFriday 13 March 2026  01:20:35 +0000 (0:00:00.126)       0:00:10.578 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set health test data] ****************************************************\nFriday 13 March 2026  01:20:36 +0000 (0:00:01.516)       0:00:12.094 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable] *************************\nFriday 13 March 2026  01:20:36 +0000 (0:00:00.300)       0:00:12.395 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if health is acceptable] *****************************\nFriday 13 March 2026  01:20:37 +0000 (0:00:00.125)       0:00:12.520 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable (strict)] ****************\nFriday 13 March 2026  01:20:37 +0000 (0:00:00.144)       0:00:12.664 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if status is OK (strict)] ****************************\nFriday 13 March 2026  01:20:37 +0000 (0:00:00.288)       0:00:12.953 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nFriday 13 March 2026  01:20:37 +0000 (0:00:00.134)       0:00:13.088 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nFriday 13 March 2026  01:20:37 +0000 (0:00:00.263)       0:00:13.351 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:20:38 +0000 (0:00:00.254)       0:00:13.606 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:20:39 +0000 (0:00:01.696)       0:00:15.302 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:20:40 +0000 (0:00:00.270)       0:00:15.572 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:40 +0000 (0:00:00.267)       0:00:15.839 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:40 +0000 (0:00:00.071)       0:00:15.911 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:20:40 +0000 (0:00:00.073)       0:00:15.984 ********** \n\nRUNNING HANDLER [Write report file] ********************************************\nFriday 13 March 2026  01:20:40 +0000 (0:00:00.075)       0:00:16.060 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:20:42 +0000 (0:00:01.459)       0:00:17.520 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-13T01:20:25+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:20:42 +0000 (0:00:00.785)       0:00:18.305 ********** \n=============================================================================== \nGet monmap info from one mon container ---------------------------------- 1.76s\nAggregate test results step one ----------------------------------------- 1.70s\nGather status data ------------------------------------------------------ 1.52s\nWrite report file ------------------------------------------------------- 1.46s\nGet container info ------------------------------------------------------ 1.22s\nCreate report output directory ------------------------------------------ 0.92s\nGet timestamp for report file ------------------------------------------- 0.81s\nPrint report file information ------------------------------------------- 0.79s\nSet quorum test data ---------------------------------------------------- 0.48s\nSet test result to passed if container is existing ---------------------- 0.47s\nSet test result to passed if ceph-mon is running ------------------------ 0.44s\nPass quorum test if all monitors are in quorum -------------------------- 0.34s\nSet fsid test vars ------------------------------------------------------ 0.30s\nPrepare test data ------------------------------------------------------- 0.30s\nSet health test data ---------------------------------------------------- 0.30s\nSet test result to failed if ceph-mon is not running -------------------- 0.30s\nPrepare test data for container existance test -------------------------- 0.29s\nFail cluster-health if health is not acceptable (strict) ---------------- 0.29s\nSet test result to failed if container is missing ----------------------- 0.29s\nAggregate test results step one ----------------------------------------- 0.28s\n+ osism validate ceph-mgrs\n\nPLAY [Ceph validate mgrs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nFriday 13 March 2026  01:21:00 +0000 (0:00:00.448)       0:00:00.448 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nFriday 13 March 2026  01:21:01 +0000 (0:00:00.816)       0:00:01.265 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nFriday 13 March 2026  01:21:02 +0000 (0:00:01.005)       0:00:02.271 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nFriday 13 March 2026  01:21:02 +0000 (0:00:00.129)       0:00:02.400 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nFriday 13 March 2026  01:21:02 +0000 (0:00:00.281)       0:00:02.681 ********** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nFriday 13 March 2026  01:21:03 +0000 (0:00:00.974)       0:00:03.656 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nFriday 13 March 2026  01:21:03 +0000 (0:00:00.278)       0:00:03.934 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:04 +0000 (0:00:00.495)       0:00:04.430 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mgr is not running] ********************\nFriday 13 March 2026  01:21:04 +0000 (0:00:00.297)       0:00:04.728 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mgr is running] ************************\nFriday 13 March 2026  01:21:04 +0000 (0:00:00.312)       0:00:05.041 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:21:05 +0000 (0:00:00.508)       0:00:05.550 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:21:05 +0000 (0:00:00.252)       0:00:05.802 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:21:05 +0000 (0:00:00.251)       0:00:06.055 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.265)       0:00:06.321 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.069)       0:00:06.390 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.079)       0:00:06.470 ********** \n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.073)       0:00:06.543 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.244)       0:00:06.788 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Define mgr module test vars] *********************************************\nFriday 13 March 2026  01:21:06 +0000 (0:00:00.246)       0:00:07.034 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather list of mgr modules] **********************************************\nFriday 13 March 2026  01:21:07 +0000 (0:00:00.144)       0:00:07.179 ********** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Parse mgr module list from json] *****************************************\nFriday 13 March 2026  01:21:09 +0000 (0:00:02.060)       0:00:09.240 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Extract list of enabled mgr modules] *************************************\nFriday 13 March 2026  01:21:09 +0000 (0:00:00.429)       0:00:09.670 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail test if mgr modules are disabled that should be enabled] ************\nFriday 13 March 2026  01:21:09 +0000 (0:00:00.303)       0:00:09.974 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass test if required mgr modules are enabled] ***************************\nFriday 13 March 2026  01:21:09 +0000 (0:00:00.139)       0:00:10.113 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nFriday 13 March 2026  01:21:10 +0000 (0:00:00.155)       0:00:10.269 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nFriday 13 March 2026  01:21:10 +0000 (0:00:00.251)       0:00:10.521 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:21:10 +0000 (0:00:00.248)       0:00:10.769 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:21:11 +0000 (0:00:01.274)       0:00:12.044 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:21:12 +0000 (0:00:00.253)       0:00:12.297 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:12 +0000 (0:00:00.243)       0:00:12.540 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:12 +0000 (0:00:00.069)       0:00:12.609 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:12 +0000 (0:00:00.067)       0:00:12.677 ********** \n\nRUNNING HANDLER [Write report file] ********************************************\nFriday 13 March 2026  01:21:12 +0000 (0:00:00.239)       0:00:12.917 ********** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:21:14 +0000 (0:00:01.264)       0:00:14.182 ********** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-13T01:21:01+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:21:14 +0000 (0:00:00.389)       0:00:14.571 ********** \n=============================================================================== \nGather list of mgr modules ---------------------------------------------- 2.06s\nAggregate test results step one ----------------------------------------- 1.28s\nWrite report file ------------------------------------------------------- 1.26s\nCreate report output directory ------------------------------------------ 1.01s\nGet container info ------------------------------------------------------ 0.97s\nGet timestamp for report file ------------------------------------------- 0.82s\nSet test result to passed if ceph-mgr is running ------------------------ 0.51s\nSet test result to passed if container is existing ---------------------- 0.50s\nParse mgr module list from json ----------------------------------------- 0.43s\nPrint report file information ------------------------------------------- 0.39s\nFlush handlers ---------------------------------------------------------- 0.38s\nSet test result to failed if ceph-mgr is not running -------------------- 0.31s\nExtract list of enabled mgr modules ------------------------------------- 0.30s\nPrepare test data ------------------------------------------------------- 0.30s\nPrepare test data for container existance test -------------------------- 0.28s\nSet test result to failed if container is missing ----------------------- 0.28s\nAggregate test results step three --------------------------------------- 0.27s\nAggregate test results step two ----------------------------------------- 0.25s\nAggregate test results step two ----------------------------------------- 0.25s\nAggregate test results step one ----------------------------------------- 0.25s\n+ osism validate ceph-osds\n\nPLAY [Ceph validate OSDs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nFriday 13 March 2026  01:21:26 +0000 (0:00:00.320)       0:00:00.320 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nFriday 13 March 2026  01:21:27 +0000 (0:00:00.678)       0:00:00.998 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nFriday 13 March 2026  01:21:27 +0000 (0:00:00.382)       0:00:01.380 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nFriday 13 March 2026  01:21:28 +0000 (0:00:00.689)       0:00:02.070 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nFriday 13 March 2026  01:21:28 +0000 (0:00:00.122)       0:00:02.193 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nFriday 13 March 2026  01:21:28 +0000 (0:00:00.125)       0:00:02.318 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nFriday 13 March 2026  01:21:28 +0000 (0:00:00.276)       0:00:02.594 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nFriday 13 March 2026  01:21:28 +0000 (0:00:00.129)       0:00:02.724 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate total number of OSDs in cluster] *******************************\nFriday 13 March 2026  01:21:29 +0000 (0:00:00.291)       0:00:03.015 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:29 +0000 (0:00:00.606)       0:00:03.621 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers on host] *********************************\nFriday 13 March 2026  01:21:30 +0000 (0:00:00.271)       0:00:03.893 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8c865601b389d1fe688c59bbce7c9c3bc0724c5899d822fc61c74068974e61fe', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b7147f1bef5ea7241362eea7f8538510c31ad51819c12903214405857b7d82ef', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4e1aa8565c9877cdf1f897b28d56ab3187247dd869e3f51bbf01bd512df70508', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '06b1cc71ccdec33e79f0b4eee83c0db488ddfd5ec1efa160310077073e5d18b3', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8e31420c6242de113d89aea4314577257644b2d30d8144daef0da155da570ba4', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '219a66a5a5bf1f0b561aeb7d424fadef452117281bb8152e3b64f5d3088edb78', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '838a854ce875d0fc141e74bf22c0a7158d3114abf1ba759e85a6e59b70c2acb1', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'f3943bcde8de3d5b646e46f400d2545dc9779a0f94c9df66985e046139e9e418', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ba676f3da3ee3f7f55f8d5410d9bc077c2fbee97d9f8442517186928e0e888b3', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a0baa27dd316ac2ef80e1cdc784cf4b694009581bd83040fb77e61232d4fb483', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': '1e74e68e4ff6794127d01d9e1b32b0fc09cbdf0c82b23241e09c044508f7496f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': '47641c347c343cbe4b73bb4d590d02d728fa115e90a0086256ecafeb5e4974de', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e5bbad3d6dce42daa045230503e1ad52f59d3e63073c9b4587768810646c39e0', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b616fd0ff3e59d7fc1c8f57d1a304fdec40e0b1c4ffde4bd698e4c9f5b5a18c1', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '006940d481062f8b2bf4b1b00067947de48a516d70b5ea5a8f48671a62d95e2c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e6c2bb5ee88ef518b998b52ae3653df452ac14b573c770bb161133d5aa820114', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '80cafc4a6dd432a2858bdd8ea6ff67fca743dd46eaac1016ae21b2b87cef727c', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '3e7ed081145c3d2b7bf8bbfe818f1b6eb0567923e5af61704515ba8d0ae9c02c', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd60cd154363827e30cbe4f4d9b5e012768419f516a1afc13c1ee5ceeb6230253', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'b95bcbea35d52acc6645634e35951aeb7a7439e77bc3e197b9e589de86441fdd', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8d3afd5541fd1475a55f8b97aef4c69e91426749be733dbd666f7163cf5de904', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ee4e96ce1c56963a8affa28df4358c450cc7d756df5439da37b5d755c8871943', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6ed0baee43bd02ab2f73e8e6b9e3dfd032583a669d7bf66b95073bf0e946e3a7', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'fe0432ecce244f4ef2fea3a7eb2166018fab34e1c6e328da7cf083287d215a44', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6e5dcda8d1bcd9cf76d0256ba123cdeaeaf3b228722799666e920c8f1a93953e', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f989112bf0b79d4606f895117eb247f2750c570247cfa7c3894e7373d20975b0', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '02f04d5ec85a7b0bde6c2e42443537c842c00aa31ecd16ccb5a70b0d90f02a84', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '56351f94e56c5920e8517dfcd0c7a9dc34f50bf8b12aca32e05a7804a6760937', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '3303bf2bba0dd445aa1761d756c623248ccfffacd119372c68568e1e104b029e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '4b1511bac231627cd4efed9cc7a0aacf8d66dde51f2448944f4aeaf81e99ce60', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'be870e67c539dee64c08627ffea04e3ac232da50d0727ac2686ef13de2a225c5', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'a80efbde99dc30df91150ce781bc04ac56c15d1ceb2a9b1bc7cbff7b99ddb4e3', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '05a44d95a05c7995789ae4bcd886fec8de74b046a87c1ceae2bcafe66782be5c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '7cf69b4487569e329087c1913e4338f59af3982ddfe9410f1cf88202b6715926', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '767104052acc36a839debca40d71b9d8c6986272d1bd940553259c82b7f82d2a', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b731b3ceaa9ba7f81628c9432a7cbcdee4b64c17d69f61494ab95e1fd973af63', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1651b8e31a45bb8d0b9f8c6e901cbb59fa9e1f6e736f41ed2f966b07b6dcb18d', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '630b5a32fbb3395f252e15063c19a5d4ee9b7864627253773df5cbf23ed00ee6', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '32eab2929619d7d147289ad2798d6157ff7ff393a74f5df88592c0ac038b95f0', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1cfaec017d1a28c40a0f888e3e6d7c67849fdc35f0c7fa0093885015957e055b', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'd713c8de9f0dc8f642f1290a61846f1c29c9a79c9ad133fc8b252485630c678b', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '86cd5d46c38aaedf25869ba48fdb9fd1763b11297da05dab58b5c654c2d949bb', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f515baadbd95b8e302c19b43f5701960b2f8a7becb4887fa5dfe39919557bb87', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c6fbb16e8cf12df44a95ab5c3d432545a320eba8e2f29507a1be7fa6dd8a64ba', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b4abc0ef2dcb6e787bd375cba32312d3e0ee01099f3a3237f4351bcbd21aa5fb', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8ddb33a2427e56dc1d44c1cf60ba5f09d310b217e628113d00601c2d73fdeb97', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '5fe0fdcc2774f76391ead83aa2df8b5a36fa8a4bf962924dd530d7c6709bd25e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': 'efebf37ff7bf9063c1f01b72f82e34857c0babb9b9d3d0bd32310d05201ae34f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fbdf31c96080ceb795826725def5e8a1b0444e30a00fe2f706b72cdb45a214ee', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'd57416ad822d57921d5d3e0ca2ad56327699af161bc334e768a6bab3ab054285', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '77597e65fb914d92c76b6c3fed3d7353f14955be3003d22a228780ca5fece769', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '52659bd4d1973760b326a0b719d38f64ec2e4fb717c5699d0da21ff4b41eddd2', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8fe38e8a78cf46bd7fd5e5d237c9de16213ea5902685c5eadcaa5c100a7fd417', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f45c4e6b4aff1e7f32694e5aed9c8b49a55076313bbc23c730b8a38a4b7a9239', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m\n\nTASK [Get count of ceph-osd containers on host] ********************************\nFriday 13 March 2026  01:21:30 +0000 (0:00:00.408)       0:00:04.301 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed when count of containers is wrong] *************\nFriday 13 March 2026  01:21:30 +0000 (0:00:00.251)       0:00:04.553 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if count matches] ******************************\nFriday 13 March 2026  01:21:31 +0000 (0:00:00.398)       0:00:04.952 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:31 +0000 (0:00:00.277)       0:00:05.230 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers that are not running] ********************\nFriday 13 March 2026  01:21:31 +0000 (0:00:00.247)       0:00:05.477 ********** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get count of ceph-osd containers that are not running] *******************\nFriday 13 March 2026  01:21:32 +0000 (0:00:00.297)       0:00:05.774 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nFriday 13 March 2026  01:21:32 +0000 (0:00:00.389)       0:00:06.163 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nFriday 13 March 2026  01:21:32 +0000 (0:00:00.262)       0:00:06.426 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if all containers are running] *****************\nFriday 13 March 2026  01:21:32 +0000 (0:00:00.242)       0:00:06.668 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:21:33 +0000 (0:00:00.258)       0:00:06.927 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:21:33 +0000 (0:00:00.501)       0:00:07.429 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:21:33 +0000 (0:00:00.219)       0:00:07.648 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.231)       0:00:07.880 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.065)       0:00:07.945 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.072)       0:00:08.017 ********** \n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.064)       0:00:08.081 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail early due to containers not running] ********************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.234)       0:00:08.316 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:34 +0000 (0:00:00.264)       0:00:08.581 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set _mon_hostname fact] **************************************************\nFriday 13 March 2026  01:21:35 +0000 (0:00:00.284)       0:00:08.865 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get ceph osd tree] *******************************************************\nFriday 13 March 2026  01:21:35 +0000 (0:00:00.682)       0:00:09.547 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Parse osd tree from JSON] ************************************************\nFriday 13 March 2026  01:21:37 +0000 (0:00:01.861)       0:00:11.409 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get OSDs that are not up or in] ******************************************\nFriday 13 March 2026  01:21:37 +0000 (0:00:00.122)       0:00:11.531 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Fail test if OSDs are not up or in] **************************************\nFriday 13 March 2026  01:21:38 +0000 (0:00:00.309)       0:00:11.840 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Pass test if OSDs are all up and in] *************************************\nFriday 13 March 2026  01:21:38 +0000 (0:00:00.125)       0:00:11.966 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:38 +0000 (0:00:00.150)       0:00:12.117 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [List ceph LVM volumes and collect data] **********************************\nFriday 13 March 2026  01:21:38 +0000 (0:00:00.274)       0:00:12.391 ********** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Parse LVM data as JSON] **************************************************\nFriday 13 March 2026  01:21:40 +0000 (0:00:02.263)       0:00:14.655 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get unencrypted and encrypted OSDs] **************************************\nFriday 13 March 2026  01:21:41 +0000 (0:00:00.488)       0:00:15.143 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of encrypted OSDs does not match] **************************\nFriday 13 March 2026  01:21:41 +0000 (0:00:00.494)       0:00:15.638 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of encrypted OSDs equals count of OSDs] ********************\nFriday 13 March 2026  01:21:42 +0000 (0:00:00.287)       0:00:15.926 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of unencrypted OSDs does not match] ************************\nFriday 13 March 2026  01:21:42 +0000 (0:00:00.450)       0:00:16.376 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************\nFriday 13 March 2026  01:21:42 +0000 (0:00:00.294)       0:00:16.671 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nFriday 13 March 2026  01:21:43 +0000 (0:00:00.292)       0:00:16.963 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get CRUSH node data of each OSD host and root node childs] ***************\nFriday 13 March 2026  01:21:43 +0000 (0:00:00.469)       0:00:17.433 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate sub test expression results] ***********************************\nFriday 13 March 2026  01:21:44 +0000 (0:00:00.712)       0:00:18.146 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail test if any sub test failed] ****************************************\nFriday 13 March 2026  01:21:44 +0000 (0:00:00.311)       0:00:18.457 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass test if no sub test failed] *****************************************\nFriday 13 March 2026  01:21:45 +0000 (0:00:00.328)       0:00:18.786 ********** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nFriday 13 March 2026  01:21:45 +0000 (0:00:00.324)       0:00:19.110 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nFriday 13 March 2026  01:21:46 +0000 (0:00:00.664)       0:00:19.775 ********** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nFriday 13 March 2026  01:21:46 +0000 (0:00:00.235)       0:00:20.010 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nFriday 13 March 2026  01:21:47 +0000 (0:00:01.511)       0:00:21.522 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nFriday 13 March 2026  01:21:47 +0000 (0:00:00.248)       0:00:21.770 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:48 +0000 (0:00:00.272)       0:00:22.043 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:48 +0000 (0:00:00.081)       0:00:22.126 ********** \n\nTASK [Flush handlers] **********************************************************\nFriday 13 March 2026  01:21:48 +0000 (0:00:00.067)       0:00:22.194 ********** \n\nRUNNING HANDLER [Write report file] ********************************************\nFriday 13 March 2026  01:21:48 +0000 (0:00:00.069)       0:00:22.263 ********** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nFriday 13 March 2026  01:21:49 +0000 (0:00:01.269)       0:00:23.533 ********** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-13T01:21:27+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:21:50 +0000 (0:00:00.576)       0:00:24.109 ********** \n=============================================================================== \nList ceph LVM volumes and collect data ---------------------------------- 2.26s\nGet ceph osd tree ------------------------------------------------------- 1.86s\nAggregate test results step one ----------------------------------------- 1.51s\nWrite report file ------------------------------------------------------- 1.27s\nGet CRUSH node data of each OSD host and root node childs --------------- 0.71s\nCreate report output directory ------------------------------------------ 0.69s\nSet _mon_hostname fact -------------------------------------------------- 0.68s\nGet timestamp for report file ------------------------------------------- 0.68s\nSet validation result to passed if no test failed ----------------------- 0.66s\nCalculate total number of OSDs in cluster ------------------------------- 0.61s\nPrint report file information ------------------------------------------- 0.58s\nAggregate test results step one ----------------------------------------- 0.50s\nGet unencrypted and encrypted OSDs -------------------------------------- 0.49s\nParse LVM data as JSON -------------------------------------------------- 0.49s\nPrepare test data ------------------------------------------------------- 0.47s\nPass if count of encrypted OSDs equals count of OSDs -------------------- 0.45s\nGet list of ceph-osd containers on host --------------------------------- 0.41s\nSet test result to failed when count of containers is wrong ------------- 0.40s\nGet count of ceph-osd containers that are not running ------------------- 0.39s\nGet extra vars for Ceph configuration ----------------------------------- 0.38s\n+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ -e /etc/redhat-release ]]\n+ source /etc/os-release\n++ PRETTY_NAME='Ubuntu 24.04.4 LTS'\n++ NAME=Ubuntu\n++ VERSION_ID=24.04\n++ VERSION='24.04.4 LTS (Noble Numbat)'\n++ VERSION_CODENAME=noble\n++ ID=ubuntu\n++ ID_LIKE=debian\n++ HOME_URL=https://www.ubuntu.com/\n++ SUPPORT_URL=https://help.ubuntu.com/\n++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/\n++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\n++ UBUNTU_CODENAME=noble\n++ LOGO=ubuntu-logo\n+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]\n+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'\n+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n\n# Status of Elasticsearch\n\n+ pushd /opt/configuration/contrib\n+ echo\n+ echo '# Status of Elasticsearch'\n+ echo\n+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s\nOK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0\n\n# Status of MariaDB\n+ echo\n+ echo '# Status of MariaDB'\n+ echo\n\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status database\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: a2f1f1b1-1e77-11f1-9a51-b241cc3da751\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: d0a8827e-1e77-11f1-a897-db3dd2836140\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.07%\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.00423789\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4818 local commits, 7019 replicated, 92 received\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 23 minutes, 15 seconds\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 149 connected, 1 running\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 209806 total, 0 slow\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 132\u001b[0m\n\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m\n\n# Status of Prometheus\n\n+ echo\n+ echo '# Status of Prometheus'\n+ echo\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy\nUnauthorized\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready\nUnauthorized\n\n# Status of RabbitMQ\n\n+ echo\n+ echo '# Status of RabbitMQ'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status messaging\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 4.1.8\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 27.3.4.1\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 212, Channels: 211, Queues: 173\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 221 total, 221 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 111/1024\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 0/0\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 4.1.8\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 27.3.4.1\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 212, Channels: 211, Queues: 173\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 221 total, 221 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.6 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 111/1024\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 0/0\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 4.1.8\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 27.3.4.1\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 212, Channels: 211, Queues: 173\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 221 total, 221 ready, 0 unacked\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 111/1024\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 0/0\u001b[0m\n\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m\n\n# Status of Redis\n\n+ echo\n+ echo '# Status of Redis'\n+ echo\n+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j\nTCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.001782s;;;0.000000;10.000000\n\n# Create backup of MariaDB database\n\n+ popd\n+ echo\n+ echo '# Create backup of MariaDB database'\n+ echo\n+ osism apply mariadb_backup -e mariadb_backup_type=full\n\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m\n\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9efa469-6931-4d78-a59c-d3d2ae5d6be5 (mariadb_backup) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f9efa469-6931-4d78-a59c-d3d2ae5d6be5 (mariadb_backup) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nFriday 13 March 2026  01:22:38 +0000 (0:00:00.168)       0:00:00.168 ********** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nFriday 13 March 2026  01:22:38 +0000 (0:00:00.329)       0:00:00.498 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nFriday 13 March 2026  01:22:39 +0000 (0:00:00.569)       0:00:01.068 ********** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nFriday 13 March 2026  01:22:39 +0000 (0:00:00.424)       0:00:01.493 ********** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Get MariaDB container facts] ***********************************\nFriday 13 March 2026  01:22:40 +0000 (0:00:00.532)       0:00:02.025 ********** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Taking full database backup via Mariabackup] *******************\nFriday 13 March 2026  01:22:43 +0000 (0:00:03.318)       0:00:05.344 ********** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nFriday 13 March 2026  01:22:59 +0000 (0:00:16.143)       0:00:21.487 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nFriday 13 March 2026  01:23:00 +0000 (0:00:00.294)       0:00:21.782 ********** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:23:00 +0000 (0:00:00.381)       0:00:22.164 ********** \n=============================================================================== \nmariadb : Taking full database backup via Mariabackup ------------------ 16.14s\nmariadb : Get MariaDB container facts ----------------------------------- 3.32s\nGroup hosts based on enabled services ----------------------------------- 0.57s\nmariadb : include_tasks ------------------------------------------------- 0.53s\nmariadb : Group MariaDB hosts based on shards --------------------------- 0.42s\nInclude mariadb post-upgrade.yml ---------------------------------------- 0.38s\nGroup hosts based on Kolla action --------------------------------------- 0.33s\nInclude mariadb post-deploy.yml ----------------------------------------- 0.29s\n+ sh -c /opt/configuration/scripts/check/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ export OS_CLOUD=admin\n+ OS_CLOUD=admin\n+ echo\n\n# OpenStack endpoints\n\n+ echo '# OpenStack endpoints'\n+ echo\n+ openstack endpoint list\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| 1caec2687c60456cbc7822a383782bc2 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |\n| 201a07ec02d14ca09fd330558a3dd6fe | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |\n| 20590df4ee8a4f86a893a053d0886293 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |\n| 2140ae4e39674fab810e3d189fabebff | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |\n| 37fead48967943f8aa6a4d759200a29e | RegionOne | cinder       | block-storage   | True    | public    | https://api.testbed.osism.xyz:8776/v3                               |\n| 394cc09f89b043938d746ffb14db3f0d | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |\n| 3ec39bdb982742749096453c50e1a2dc | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |\n| 51ae298bdeab4cfaab67e021b474bb2b | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |\n| 571ffbcfe3724f59972ee4b86febf04c | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |\n| 7804c9d6b7384114bbd1ab1424c980f8 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |\n| 7e356053703344fabf9de43c4a88ad33 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |\n| 81bcdaf0ddda4ae993189e130798c6e5 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |\n| 985d48b634c2474aae6827f3508edab8 | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |\n| 9bd56d8861d74b6690965436ba896a33 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |\n| a29638ae2912484cae5a2933c7f96f95 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |\n| b498e434c068462f903caa8e515293f5 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |\n| b96bff7722bc4d6ea39ed79e83829c57 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |\n| d20fbe52d0424575b1f01a1ef481fbc6 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |\n| d64f2d9293dc4fbb896e915c9e23aca7 | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |\n| dcc95226d4eb4c32a28c7c55ead33887 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |\n| e415e18e83734c209dd66ba9d74eb6be | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |\n| e94e8394ec4e4a7a98101bfca6c4ee20 | RegionOne | cinder       | block-storage   | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3                           |\n| f5d9ad80054d4ff7a86d8175b186ef9b | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |\n| fa7634d91e82403c87361c99d3d62137 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n\n# Cinder\n\n+ echo\n+ echo '# Cinder'\n+ echo\n+ openstack volume service list\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| Binary           | Host                       | Zone     | Status  | State | Updated At                 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-13T01:22:58.000000 |\n| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-13T01:22:59.000000 |\n| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-13T01:22:58.000000 |\n| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:22:59.000000 |\n| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:23:04.000000 |\n| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:23:05.000000 |\n| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-13T01:23:03.000000 |\n| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-13T01:23:05.000000 |\n| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-13T01:23:06.000000 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n\n# Neutron\n\n+ echo\n+ echo '# Neutron'\n+ echo\n+ openstack network agent list\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |\n| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n+ openstack network service provider list\n+---------------+------+---------+\n| Service Type  | Name | Default |\n+---------------+------+---------+\n| L3_ROUTER_NAT | ovn  | True    |\n+---------------+------+---------+\n\n# Nova\n\n+ echo\n+ echo '# Nova'\n+ echo\n+ openstack compute service list\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| 34696851-f8b5-44c1-abf2-1235a653e4ab | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |\n| f87f6b66-58f0-453b-b277-6f795b5e975b | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |\n| 408149be-f79f-42b0-b950-5da3ddcb592b | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |\n| 2b210f2d-6743-4ed5-9797-fe8a1fe9dda5 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-13T01:23:07.000000 |\n| e28b9961-0e3a-441a-b16e-c2e70cd3432c | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-13T01:23:08.000000 |\n| b7b61d6d-742b-4109-b14d-dcbe0ca96e1c | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-13T01:23:08.000000 |\n| 39372ed5-73ea-46d1-bec2-b9d98d410998 | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-13T01:23:11.000000 |\n| babb32bb-811a-4abc-857e-666a13237230 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-13T01:23:11.000000 |\n| 7f4b5cac-2b39-4e9e-9df5-5fd5e72908b0 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-13T01:23:12.000000 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n+ openstack hypervisor list\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| 36ff5121-2fb8-42f3-8f4f-36344af1223b | testbed-node-3      | QEMU            | 192.168.16.13 | up    |\n| b4594c09-21eb-4c9a-9400-487c30120c44 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |\n| 0dccbc3a-9367-460d-ae63-5bb2a224353c | testbed-node-5      | QEMU            | 192.168.16.15 | up    |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n+ echo\n\n# Run OpenStack test play\n\n+ echo '# Run OpenStack test play'\n+ echo\n+ osism apply --environment openstack test\n\u001b[32m2026-03-13 01:23:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m\n\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m\n\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff00ddd-7547-4db0-9f4c-3644ccf273b4 (test) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cff00ddd-7547-4db0-9f4c-3644ccf273b4 (test) has been started and output is visible here.\u001b[0m\n\nPLAY [Create test project] *****************************************************\n\nTASK [Create test domain] ******************************************************\nFriday 13 March 2026  01:23:32 +0000 (0:00:00.051)       0:00:00.051 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test-admin user] **************************************************\nFriday 13 March 2026  01:23:36 +0000 (0:00:03.268)       0:00:03.320 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add manager role to user test-admin] *************************************\nFriday 13 March 2026  01:23:39 +0000 (0:00:03.710)       0:00:07.030 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test project] *****************************************************\nFriday 13 March 2026  01:23:46 +0000 (0:00:06.559)       0:00:13.590 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test user] ********************************************************\nFriday 13 March 2026  01:23:50 +0000 (0:00:04.045)       0:00:17.635 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add member roles to user test] *******************************************\nFriday 13 March 2026  01:23:54 +0000 (0:00:04.186)       0:00:21.822 ********** \n\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m\n\nTASK [Create test server group] ************************************************\nFriday 13 March 2026  01:24:05 +0000 (0:00:11.380)       0:00:33.202 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create ssh security group] ***********************************************\nFriday 13 March 2026  01:24:10 +0000 (0:00:04.789)       0:00:37.992 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to ssh security group] ******************************************\nFriday 13 March 2026  01:24:14 +0000 (0:00:04.161)       0:00:42.153 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create icmp security group] **********************************************\nFriday 13 March 2026  01:24:18 +0000 (0:00:03.831)       0:00:45.984 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to icmp security group] *****************************************\nFriday 13 March 2026  01:24:22 +0000 (0:00:03.637)       0:00:49.622 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test keypair] *****************************************************\nFriday 13 March 2026  01:24:26 +0000 (0:00:03.851)       0:00:53.474 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test network] *****************************************************\nFriday 13 March 2026  01:24:29 +0000 (0:00:03.798)       0:00:57.273 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test subnet] ******************************************************\nFriday 13 March 2026  01:24:34 +0000 (0:00:04.801)       0:01:02.074 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test router] ******************************************************\nFriday 13 March 2026  01:24:40 +0000 (0:00:05.719)       0:01:07.794 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Manage test instances and volumes] ***************************************\n\nTASK [Get test server group] ***************************************************\nFriday 13 March 2026  01:24:51 +0000 (0:00:10.649)       0:01:18.444 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Detach test volume] ******************************************************\nFriday 13 March 2026  01:24:54 +0000 (0:00:03.414)       0:01:21.858 ********** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test volume] ******************************************************\nFriday 13 March 2026  01:24:54 +0000 (0:00:00.053)       0:01:21.911 ********** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test instances] ***************************************************\nFriday 13 March 2026  01:24:54 +0000 (0:00:00.039)       0:01:21.951 ********** \n\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m\n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Wait for instance deletion to complete] **********************************\nFriday 13 March 2026  01:24:54 +0000 (0:00:00.158)       0:01:22.110 ********** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Create test instances] ***************************************************\nFriday 13 March 2026  01:24:54 +0000 (0:00:00.138)       0:01:22.249 ********** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for instance creation to complete] **********************************\nFriday 13 March 2026  01:24:59 +0000 (0:00:04.635)       0:01:26.885 ********** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j430084601681.2616', 'results_file': '/ansible/.ansible_async/j430084601681.2616', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j504665637972.2641', 'results_file': '/ansible/.ansible_async/j504665637972.2641', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j984402392910.2666', 'results_file': '/ansible/.ansible_async/j984402392910.2666', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j508465868329.2691', 'results_file': '/ansible/.ansible_async/j508465868329.2691', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j65206343094.2716', 'results_file': '/ansible/.ansible_async/j65206343094.2716', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add metadata to instances] ***********************************************\nFriday 13 March 2026  01:25:57 +0000 (0:00:57.655)       0:02:24.540 ********** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for metadata to be added] *******************************************\nFriday 13 March 2026  01:26:01 +0000 (0:00:04.524)       0:02:29.065 ********** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j154185062180.2827', 'results_file': '/ansible/.ansible_async/j154185062180.2827', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j218872645057.2852', 'results_file': '/ansible/.ansible_async/j218872645057.2852', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j220671646960.2877', 'results_file': '/ansible/.ansible_async/j220671646960.2877', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j915736299113.2902', 'results_file': '/ansible/.ansible_async/j915736299113.2902', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j294611988658.2927', 'results_file': '/ansible/.ansible_async/j294611988658.2927', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add tag to instances] ****************************************************\nFriday 13 March 2026  01:26:11 +0000 (0:00:09.628)       0:02:38.693 ********** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for tags to be added] ***********************************************\nFriday 13 March 2026  01:26:15 +0000 (0:00:04.182)       0:02:42.875 ********** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j966377388877.3003', 'results_file': '/ansible/.ansible_async/j966377388877.3003', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j927214934809.3028', 'results_file': '/ansible/.ansible_async/j927214934809.3028', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j964121494158.3054', 'results_file': '/ansible/.ansible_async/j964121494158.3054', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j287701880302.3080', 'results_file': '/ansible/.ansible_async/j287701880302.3080', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j272798000334.3106', 'results_file': '/ansible/.ansible_async/j272798000334.3106', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Create test volume] ******************************************************\nFriday 13 March 2026  01:26:25 +0000 (0:00:09.488)       0:02:52.364 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Attach test volume] ******************************************************\nFriday 13 March 2026  01:26:31 +0000 (0:00:06.542)       0:02:58.907 ********** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create floating ip address] **********************************************\nFriday 13 March 2026  01:26:45 +0000 (0:00:13.454)       0:03:12.361 ********** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Print floating ip address] ***********************************************\nFriday 13 March 2026  01:26:50 +0000 (0:00:04.999)       0:03:17.361 ********** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"192.168.112.161\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:26:50 +0000 (0:00:00.042)       0:03:17.403 ********** \n=============================================================================== \nWait for instance creation to complete --------------------------------- 57.66s\nAttach test volume ----------------------------------------------------- 13.45s\nAdd member roles to user test ------------------------------------------ 11.38s\nCreate test router ----------------------------------------------------- 10.65s\nWait for metadata to be added ------------------------------------------- 9.63s\nWait for tags to be added ----------------------------------------------- 9.49s\nAdd manager role to user test-admin ------------------------------------- 6.56s\nCreate test volume ------------------------------------------------------ 6.54s\nCreate test subnet ------------------------------------------------------ 5.72s\nCreate floating ip address ---------------------------------------------- 5.00s\nCreate test network ----------------------------------------------------- 4.80s\nCreate test server group ------------------------------------------------ 4.79s\nCreate test instances --------------------------------------------------- 4.64s\nAdd metadata to instances ----------------------------------------------- 4.53s\nCreate test user -------------------------------------------------------- 4.19s\nAdd tag to instances ---------------------------------------------------- 4.18s\nCreate ssh security group ----------------------------------------------- 4.16s\nCreate test project ----------------------------------------------------- 4.05s\nAdd rule to icmp security group ----------------------------------------- 3.85s\nAdd rule to ssh security group ------------------------------------------ 3.83s\n+ server_list\n+ openstack --os-cloud test server list\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE | test=192.168.112.141, 192.168.200.30  | N/A (booted from volume) | SCS-1L-1 |\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE | test=192.168.112.156, 192.168.200.80  | N/A (booted from volume) | SCS-1L-1 |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE | test=192.168.112.123, 192.168.200.121 | N/A (booted from volume) | SCS-1L-1 |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE | test=192.168.112.191, 192.168.200.13  | N/A (booted from volume) | SCS-1L-1 |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE | test=192.168.112.161, 192.168.200.236 | N/A (booted from volume) | SCS-1L-1 |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n+ openstack --os-cloud test server show test\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-13T01:25:30.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.161, 192.168.200.236                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-13T01:25:03Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | e6988b32-498d-46d2-b1fe-3cea242e8f3e                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-13T01:26:02Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='3857ac78-c588-4e7e-ba0a-80a8215a033b'                                                                                                                                                                                                                                                                                                             |\n|                                     | delete_on_termination='False', id='30b2a964-3736-450d-84ca-79a85f4022cb'                                                                                                                                                                                                                                                                                                            |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-1\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-13T01:25:31.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.191, 192.168.200.13                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-13T01:25:03Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 46752925-b128-4628-8a08-dc074af15a37                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-13T01:26:03Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='2ade6197-02cf-46f9-bdd3-8730978e79f1'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-2\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-13T01:25:33.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.156, 192.168.200.80                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-13T01:25:05Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 488c8c86-dc7a-4826-9b1b-9abaf28c3d20                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='b36e9ede-44fb-447c-80dc-63b604527332'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-3\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-13T01:25:31.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.123, 192.168.200.121                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-13T01:25:05Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 957c72c6-9401-4f8f-877a-087c45362881                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='db590097-db33-4136-8b60-1bc92af56e8e'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-4\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-03-13T01:25:34.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.141, 192.168.200.30                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-03-13T01:25:08Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | ea1c861b1c235aaadb8497ab6289fa1013a500fd3b8947c250900d9a                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 27cf0c91-b78c-4057-8905-5e463648a2a1                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-03-13T01:26:05Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='6bfa19e9-07ae-4e96-9435-4be3397815d8'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ server_ping\n++ tr -d '\\r'\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.123\nPING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.\n64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=5.69 ms\n64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.36 ms\n64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=2.07 ms\n\n--- 192.168.112.123 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 2.068/3.371/5.691/1.644 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.161\nPING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.\n64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=10.6 ms\n64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=1.50 ms\n64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.18 ms\n\n--- 192.168.112.161 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.178/4.422/10.586/4.360 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=4.26 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.18 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.93 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.934/2.792/4.259/1.042 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.44 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.45 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.69 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.694/3.194/5.441/1.618 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.191\nPING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.\n64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=4.89 ms\n64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.18 ms\n64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.80 ms\n\n--- 192.168.112.191 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.800/2.956/4.887/1.373 ms\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4\n\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:27:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m\n\u001b[32m2026-03-13 01:27:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:27:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:27:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:27:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:27:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:28:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m\n\u001b[32m2026-03-13 01:28:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m\n\u001b[32m2026-03-13 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:28:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:29:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m\n\u001b[32m2026-03-13 01:29:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:29:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ tr -d '\\r'\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.123\nPING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.\n64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=4.71 ms\n64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.52 ms\n64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.74 ms\n\n--- 192.168.112.123 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.741/2.989/4.708/1.255 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.161\nPING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.\n64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=10.3 ms\n64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.67 ms\n64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.71 ms\n\n--- 192.168.112.161 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.705/4.906/10.349/3.868 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.18 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.24 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.53 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.527/3.979/8.175/2.981 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.07 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.60 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.71 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.707/3.126/5.074/1.424 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.191\nPING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.\n64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=6.09 ms\n64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.05 ms\n64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.85 ms\n\n--- 192.168.112.191 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.854/3.331/6.093/1.954 ms\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5\n\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:30:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo migratable instances found on node testbed-node-5\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.123\nPING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.\n64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=9.67 ms\n64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.27 ms\n64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.56 ms\n\n--- 192.168.112.123 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.561/4.501/9.668/3.665 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.161\nPING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.\n64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=6.14 ms\n64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.72 ms\n64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.72 ms\n\n--- 192.168.112.161 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.716/3.524/6.141/1.894 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=9.46 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=1.96 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.83 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.830/4.415/9.460/3.567 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.54 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.76 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.25 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.247/2.848/5.535/1.911 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.191\nPING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.\n64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=3.07 ms\n64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.14 ms\n64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.60 ms\n\n--- 192.168.112.191 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.598/2.270/3.073/0.609 ms\n+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3\n\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:30:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 27cf0c91-b78c-4057-8905-5e463648a2a1\u001b[0m\n\u001b[32m2026-03-13 01:30:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:30:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:30:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m\n\u001b[32m2026-03-13 01:31:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m\n\u001b[32m2026-03-13 01:31:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:31:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:32:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m\n\u001b[32m2026-03-13 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m\n\u001b[32m2026-03-13 01:32:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:32:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.123\nPING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.\n64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=5.53 ms\n64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.54 ms\n64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.66 ms\n\n--- 192.168.112.123 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.661/3.242/5.527/1.654 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.161\nPING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.\n64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=9.38 ms\n64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=3.00 ms\n64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.84 ms\n\n--- 192.168.112.161 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.841/4.741/9.379/3.313 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.82 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=1.92 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.97 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.922/4.238/8.820/3.239 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.29 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.77 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.68 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.681/2.915/5.290/1.679 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.191\nPING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.\n64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=4.94 ms\n64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=1.97 ms\n64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.53 ms\n\n--- 192.168.112.191 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.530/2.815/4.942/1.514 ms\n+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4\n\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-03-13 01:33:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 27cf0c91-b78c-4057-8905-5e463648a2a1\u001b[0m\n\u001b[32m2026-03-13 01:33:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:33:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:34:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m\n\u001b[32m2026-03-13 01:34:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m\n\u001b[32m2026-03-13 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:35:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m\n\u001b[32m2026-03-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-03-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m\n\u001b[32m2026-03-13 01:36:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m\n\u001b[32m2026-03-13 01:36:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |\n| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |\n| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |\n| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |\n| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.123\nPING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.\n64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=9.30 ms\n64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.28 ms\n64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=2.07 ms\n\n--- 192.168.112.123 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 2.069/4.549/9.297/3.358 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.161\nPING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.\n64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=8.07 ms\n64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.19 ms\n64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.75 ms\n\n--- 192.168.112.161 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.746/4.001/8.074/2.885 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.156\nPING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.\n64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=4.63 ms\n64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.12 ms\n64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.60 ms\n\n--- 192.168.112.156 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.604/2.784/4.634/1.324 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.77 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.75 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.70 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.697/3.071/5.767/1.906 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.191\nPING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.\n64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=5.28 ms\n64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.14 ms\n64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.74 ms\n\n--- 192.168.112.191 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.740/3.051/5.278/1.582 ms",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# CHECK",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# CHECK'",
                                "+ echo",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-manager",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager ps",
                                "\u001b[32m2026-03-13 01:19:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m",
                                "CONTAINER ID   IMAGE                                                           COMMAND                  CREATED          STATUS                    PORTS                                                                             NAMES",
                                "6c9c3f91c326   registry.osism.tech/kolla/prometheus-blackbox-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_blackbox_exporter",
                                "808be0674e46   registry.osism.tech/kolla/prometheus-alertmanager:2025.1        \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_alertmanager",
                                "df0c79486a88   registry.osism.tech/kolla/prometheus-cadvisor:2025.1            \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                                                                                               prometheus_cadvisor",
                                "0bf2b9aedfe6   registry.osism.tech/kolla/prometheus-node-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                                                                                               prometheus_node_exporter",
                                "72fc99802dae   registry.osism.tech/kolla/prometheus-server:2025.1              \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                                                                                               prometheus_server",
                                "0487d415e717   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   18 minutes ago   Up 17 minutes                                                                                               cephclient",
                                "d6aa562df796   registry.osism.tech/kolla/cron:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               cron",
                                "0111d7689ed8   registry.osism.tech/kolla/kolla-toolbox:2025.1                  \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                                                                                               kolla_toolbox",
                                "0b2f1a7a6a02   registry.osism.tech/kolla/fluentd:2025.1                        \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                                                                                               fluentd",
                                "acfaccc59dd1   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   30 minutes ago   Up 29 minutes (healthy)   80/tcp                                                                            phpmyadmin",
                                "cc26c373effc   registry.osism.tech/osism/openstackclient:2025.1                \"/usr/bin/dumb-init \u2026\"   31 minutes ago   Up 30 minutes                                                                                               openstackclient",
                                "a0520e08f520   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   31 minutes ago   Up 30 minutes (healthy)   8080/tcp                                                                          homer",
                                "10e81a04583b   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   54 minutes ago   Up 53 minutes (healthy)   192.168.16.5:3128->3128/tcp                                                       squid",
                                "faeb53a8dc03   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     manager-inventory_reconciler-1",
                                "ec56de84559b   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     osism-kubernetes",
                                "913c2ab24057   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     osism-ansible",
                                "8a07529d0df1   registry.osism.tech/osism/kolla-ansible:2025.1                  \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     kolla-ansible",
                                "06851db83dd9   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   58 minutes ago   Up 37 minutes (healthy)                                                                                     ceph-ansible",
                                "3d43f54011a4   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   58 minutes ago   Up 38 minutes (healthy)   8000/tcp                                                                          manager-ara-server-1",
                                "093c1ecb28f9   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes (healthy)   6379/tcp                                                                          manager-redis-1",
                                "1903eeaddbd6   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-flower-1",
                                "9c3b1270a042   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes (healthy)   3306/tcp                                                                          manager-mariadb-1",
                                "1fca8a62a0b5   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-openstack-1",
                                "2c059a1d2a71   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-beat-1",
                                "f924682c1247   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     manager-listener-1",
                                "d58e9b803ce9   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   58 minutes ago   Up 38 minutes             192.168.16.5:3000->3000/tcp                                                       osism-frontend",
                                "70ff0fbdeaa8   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   58 minutes ago   Up 38 minutes (healthy)   192.168.16.5:8000->8000/tcp                                                       manager-api-1",
                                "caf640858b6b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   58 minutes ago   Up 38 minutes (healthy)                                                                                     osismclient",
                                "afba45b5e55b   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   59 minutes ago   Up 59 minutes (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik",
                                "",
                                "## Images @ testbed-manager",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager images",
                                "REPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE",
                                "registry.osism.tech/osism/osism-ansible                  latest           3c82227097a8   About an hour ago   613MB",
                                "registry.osism.tech/osism/kolla-ansible                  2025.1           385f40f5bb87   About an hour ago   610MB",
                                "registry.osism.tech/osism/osism                          latest           3376395f1aa5   About an hour ago   406MB",
                                "registry.osism.tech/osism/ceph-ansible                   reef             c0b375cdf34b   About an hour ago   560MB",
                                "registry.osism.tech/osism/osism-kubernetes               latest           bb68cd12c222   About an hour ago   1.22GB",
                                "registry.osism.tech/osism/osism-frontend                 latest           2a698a7d7ce5   About an hour ago   232MB",
                                "registry.osism.tech/osism/inventory-reconciler           latest           0dc880827435   About an hour ago   335MB",
                                "registry.osism.tech/osism/openstackclient                2025.1           26b681b56b01   22 hours ago        212MB",
                                "registry.osism.tech/osism/cephclient                     reef             19951ee12217   22 hours ago        453MB",
                                "registry.osism.tech/kolla/kolla-toolbox                  2025.1           a7d5f8f5723c   23 hours ago        677MB",
                                "registry.osism.tech/kolla/fluentd                        2025.1           57639eec9828   23 hours ago        585MB",
                                "registry.osism.tech/kolla/cron                           2025.1           69f3c33abd1e   23 hours ago        271MB",
                                "registry.osism.tech/kolla/prometheus-server              2025.1           cdeb71ca24f7   23 hours ago        855MB",
                                "registry.osism.tech/kolla/prometheus-blackbox-exporter   2025.1           2a08c0f63ea9   23 hours ago        314MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor            2025.1           98f129deee0e   23 hours ago        363MB",
                                "registry.osism.tech/kolla/prometheus-alertmanager        2025.1           6d47b066a3eb   23 hours ago        409MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter       2025.1           ce49e82a896f   23 hours ago        311MB",
                                "registry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   6 weeks ago         41.4MB",
                                "registry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   3 months ago        11.5MB",
                                "registry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   3 months ago        334MB",
                                "phpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   5 months ago        742MB",
                                "registry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB",
                                "registry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   7 months ago        226MB",
                                "registry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   21 months ago       146MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-0",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "d7a63897fdbd   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "a7158a6cc794   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "1208e6186247   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "407e2d2c5e67   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "071c7d07fc86   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "3fa07550e17b   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   6 minutes ago    Up 6 minutes (healthy)              nova_novncproxy",
                                "02a2b2aa767b   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "c3e9eebc5b69   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata",
                                "cc4cf05d3c43   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api",
                                "edfb39ff2955   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "49ea21131d32   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "65c8f3fa6eb4   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes                       grafana",
                                "5d7214f53bdd   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "cb4bc32826a9   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "70870ae1ee11   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "dd0afc8356d0   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "f5500f3411e3   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter",
                                "4f542e28f74c   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_cadvisor",
                                "944b8b59133b   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "eab91fc6e472   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "ec67d88396ea   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter",
                                "871651d1bc43   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor",
                                "6c08ddc7ac34   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api",
                                "234d84c97c56   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server",
                                "4705fcd70466   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api",
                                "4018c9f2e667   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker",
                                "37ceb648b4ec   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns",
                                "5bd0f24ccb91   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer",
                                "3a8b6f299bfe   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central",
                                "482b42f701c6   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api",
                                "8a80fea66d1f   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "4c3cf114c1f4   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker",
                                "eb1031c2c652   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_keystone_listener",
                                "dc2ebb5b50c1   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api",
                                "ad46f532a250   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-0",
                                "a395f068dba4   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "b0ef065ac852   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "8190cbadd291   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "4e6e6a06584a   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   19 minutes ago   Up 19 minutes (healthy)             horizon",
                                "15822dfe46e3   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   20 minutes ago   Up 20 minutes (healthy)             mariadb",
                                "2fcf94ade88a   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards",
                                "445c7143fd80   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch",
                                "64a2621e7382   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-0",
                                "c6bf31483a83   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "66bad52d2efe   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "8aef63697d89   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "0bf63e175326   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd",
                                "055dec7c911d   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1",
                                "e7a60857f864   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_sb_db",
                                "27298373da9d   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_nb_db",
                                "e9e57f170ae9   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "8f8762892bf0   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-0",
                                "709432bb4858   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             rabbitmq",
                                "0a94b45018f0   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "20a365466b20   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_db",
                                "d1647407d1e9   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "8b18a340cf76   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "56aacb432599   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "cf316d177793   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "6538d643b237   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       kolla_toolbox",
                                "e47483949a18   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-0",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB",
                                "registry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB",
                                "registry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB",
                                "registry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB",
                                "registry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB",
                                "registry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB",
                                "registry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB",
                                "registry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB",
                                "registry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB",
                                "registry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB",
                                "registry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB",
                                "registry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB",
                                "registry.osism.tech/kolla/ceilometer-notification             2025.1    ece9531359a7   23 hours ago   990MB",
                                "registry.osism.tech/kolla/ceilometer-central                  2025.1    85a92ad61831   23 hours ago   991MB",
                                "registry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB",
                                "registry.osism.tech/kolla/aodh-api                            2025.1    cbdf1972f1eb   23 hours ago   988MB",
                                "registry.osism.tech/kolla/aodh-evaluator                      2025.1    a958cce06a38   23 hours ago   989MB",
                                "registry.osism.tech/kolla/aodh-notifier                       2025.1    0c3a6cf5d0cf   23 hours ago   989MB",
                                "registry.osism.tech/kolla/aodh-listener                       2025.1    e9a41838f69b   23 hours ago   989MB",
                                "registry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB",
                                "registry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB",
                                "registry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB",
                                "registry.osism.tech/kolla/skyline-console                     2025.1    ebd12ce3748e   23 hours ago   1.06GB",
                                "registry.osism.tech/kolla/skyline-apiserver                   2025.1    d463dff27fc9   23 hours ago   1.01GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB",
                                "registry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB",
                                "registry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB",
                                "registry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-1",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "3a0122e25fb1   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "5dfe2326c2b0   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "024ef5c3f0d4   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "013ec66becca   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "f61f87918eb2   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "9194f09ba2d8   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   6 minutes ago    Up 6 minutes (healthy)              nova_novncproxy",
                                "4a481e4b3739   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "6b60b417b766   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata",
                                "940a90b42bac   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api",
                                "4ecfee1e6c47   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "8acfc72e80bf   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana",
                                "0dd4b999a79c   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "a406a59c5bd8   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "3255c40a34e1   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "7bab09b9da97   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "4e16feaea168   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "fd9991dd1999   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter",
                                "517d8d616064   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor",
                                "f072ca386a63   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "3d599f288bee   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "0ffd8d0b7ef9   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter",
                                "241b36fba680   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor",
                                "b1736d9cea81   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api",
                                "e38e68095a28   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server",
                                "9451dc358496   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api",
                                "3b79dea4bbda   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker",
                                "1182c997c693   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns",
                                "17849d7f8f56   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer",
                                "f95bb464161d   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central",
                                "7fcf8a972a49   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api",
                                "a56c14220e7e   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "25548567dfff   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             barbican_worker",
                                "71790b96b08e   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_keystone_listener",
                                "1c9abb6281d6   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api",
                                "b28d19404e6b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-1",
                                "46caf03bfa65   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "79fea1e636a9   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "5a69b4644354   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon",
                                "5df73378b3f8   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "0859ecf8b321   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards",
                                "6f5c4963924d   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb",
                                "7d7816f04e61   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 21 minutes (healthy)             opensearch",
                                "c922c44eca4b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-1",
                                "7b9b7c77f504   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "1327ecce99d2   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "0fe32ee24f09   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "f525617ce17e   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd",
                                "da9b94edc26c   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq",
                                "be63fc264077   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1",
                                "dfbe085cd6fa   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db",
                                "00e07a8211f2   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_nb_db",
                                "b99614e57c6e   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "cb0a6fe177a7   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   27 minutes ago   Up 27 minutes                       ceph-mon-testbed-node-1",
                                "4c6ae9cc1d7d   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "94f1b5442875   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_db",
                                "b3670fa142db   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "6e430df82abd   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "da2d00ec0ae4   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "b2ce6bac3ad5   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "8a1e8d661817   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox",
                                "ee7b7af32b6c   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-1",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB",
                                "registry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB",
                                "registry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB",
                                "registry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB",
                                "registry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB",
                                "registry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB",
                                "registry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB",
                                "registry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB",
                                "registry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB",
                                "registry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB",
                                "registry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB",
                                "registry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB",
                                "registry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB",
                                "registry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB",
                                "registry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB",
                                "registry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB",
                                "registry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB",
                                "registry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB",
                                "registry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB",
                                "registry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB",
                                "registry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "",
                                "## Containers @ testbed-node-2",
                                "",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "e6817ae3c070   registry.osism.tech/kolla/octavia-worker:2025.1                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "cd5f49b10bcd   registry.osism.tech/kolla/octavia-housekeeping:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_housekeeping",
                                "344287924d1f   registry.osism.tech/kolla/octavia-health-manager:2025.1              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "8462bbab99fd   registry.osism.tech/kolla/octavia-driver-agent:2025.1                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "f7ed50e200d2   registry.osism.tech/kolla/octavia-api:2025.1                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "372277f6f27b   registry.osism.tech/kolla/nova-novncproxy:2025.1                     \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_novncproxy",
                                "a895ff617c8b   registry.osism.tech/kolla/nova-conductor:2025.1                      \"dumb-init --single-\u2026\"   7 minutes ago    Up 7 minutes (healthy)              nova_conductor",
                                "dd49ed6635c6   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_metadata",
                                "9097404d4703   registry.osism.tech/kolla/nova-api:2025.1                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api",
                                "0efb91e22778   registry.osism.tech/kolla/nova-scheduler:2025.1                      \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_scheduler",
                                "222776286f9d   registry.osism.tech/kolla/grafana:2025.1                             \"dumb-init --single-\u2026\"   11 minutes ago   Up 10 minutes                       grafana",
                                "872c63553423   registry.osism.tech/kolla/cinder-backup:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "b532d81835ac   registry.osism.tech/kolla/cinder-volume:2025.1                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "6fd57c357f33   registry.osism.tech/kolla/cinder-scheduler:2025.1                    \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_scheduler",
                                "e40ebd44e115   registry.osism.tech/kolla/cinder-api:2025.1                          \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_api",
                                "585e27543ffa   registry.osism.tech/kolla/glance-api:2025.1                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "224c4450d4d9   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2025.1   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter",
                                "7185f150b760   registry.osism.tech/kolla/prometheus-cadvisor:2025.1                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor",
                                "df0ed0ec176a   registry.osism.tech/kolla/prometheus-memcached-exporter:2025.1       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "ac935971b7fe   registry.osism.tech/kolla/prometheus-mysqld-exporter:2025.1          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "6569728c8720   registry.osism.tech/kolla/prometheus-node-exporter:2025.1            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter",
                                "7610a3e2d08b   registry.osism.tech/kolla/magnum-conductor:2025.1                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor",
                                "d229ebe45517   registry.osism.tech/kolla/magnum-api:2025.1                          \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_api",
                                "ff3788c1b8b8   registry.osism.tech/kolla/neutron-server:2025.1                      \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             neutron_server",
                                "2722ff7750bc   registry.osism.tech/kolla/placement-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             placement_api",
                                "ceea4d26aa1f   registry.osism.tech/kolla/designate-worker:2025.1                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_worker",
                                "5a28a4352cb4   registry.osism.tech/kolla/designate-mdns:2025.1                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_mdns",
                                "b414321f3049   registry.osism.tech/kolla/designate-producer:2025.1                  \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_producer",
                                "f794b9f6ff79   registry.osism.tech/kolla/designate-central:2025.1                   \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_central",
                                "609bba6b2a0f   registry.osism.tech/kolla/designate-api:2025.1                       \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_api",
                                "364c282522f8   registry.osism.tech/kolla/designate-backend-bind9:2025.1             \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             designate_backend_bind9",
                                "ebe73c91b961   registry.osism.tech/kolla/barbican-worker:2025.1                     \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_worker",
                                "571859c31679   registry.osism.tech/kolla/barbican-keystone-listener:2025.1          \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_keystone_listener",
                                "15af3a2eeaa5   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   16 minutes ago   Up 16 minutes                       ceph-mgr-testbed-node-2",
                                "62b6bb95b695   registry.osism.tech/kolla/barbican-api:2025.1                        \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             barbican_api",
                                "f2b97b58356e   registry.osism.tech/kolla/keystone:2025.1                            \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone",
                                "98954b0e8036   registry.osism.tech/kolla/keystone-fernet:2025.1                     \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_fernet",
                                "f21f0479608c   registry.osism.tech/kolla/horizon:2025.1                             \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             horizon",
                                "53b872fd5cdf   registry.osism.tech/kolla/keystone-ssh:2025.1                        \"dumb-init --single-\u2026\"   18 minutes ago   Up 18 minutes (healthy)             keystone_ssh",
                                "0ece719a6c89   registry.osism.tech/kolla/mariadb-server:2025.1                      \"dumb-init -- kolla_\u2026\"   21 minutes ago   Up 21 minutes (healthy)             mariadb",
                                "5e491dcd59f9   registry.osism.tech/kolla/opensearch-dashboards:2025.1               \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             opensearch_dashboards",
                                "b055fca7034b   registry.osism.tech/kolla/opensearch:2025.1                          \"dumb-init --single-\u2026\"   22 minutes ago   Up 22 minutes (healthy)             opensearch",
                                "1479c49abc78   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    23 minutes ago   Up 23 minutes                       ceph-crash-testbed-node-2",
                                "7a772282ae0b   registry.osism.tech/kolla/keepalived:2025.1                          \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes                       keepalived",
                                "abdf42e8302a   registry.osism.tech/kolla/proxysql:2025.1                            \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             proxysql",
                                "0a869a760f28   registry.osism.tech/kolla/haproxy:2025.1                             \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             haproxy",
                                "be42f02e5b3c   registry.osism.tech/kolla/rabbitmq:2025.1                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             rabbitmq",
                                "deb57168100c   registry.osism.tech/kolla/ovn-northd:2025.1                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       ovn_northd",
                                "d4052e508b85   registry.osism.tech/kolla/ovn-sb-db-relay:2025.1                     \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db_relay_1",
                                "0028e2095a55   registry.osism.tech/kolla/ovn-sb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_sb_db",
                                "de0d08a3fa58   registry.osism.tech/kolla/ovn-nb-db-server:2025.1                    \"dumb-init --single-\u2026\"   26 minutes ago   Up 25 minutes                       ovn_nb_db",
                                "40f00350cb17   registry.osism.tech/kolla/ovn-controller:2025.1                      \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes                       ovn_controller",
                                "73131559e39b   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   28 minutes ago   Up 28 minutes                       ceph-mon-testbed-node-2",
                                "ffcfd9858ab9   registry.osism.tech/kolla/openvswitch-vswitchd:2025.1                \"dumb-init --single-\u2026\"   28 minutes ago   Up 28 minutes (healthy)             openvswitch_vswitchd",
                                "3801462a5332   registry.osism.tech/kolla/openvswitch-db-server:2025.1               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             openvswitch_db",
                                "2785382df2cf   registry.osism.tech/kolla/redis-sentinel:2025.1                      \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis_sentinel",
                                "686809c29d19   registry.osism.tech/kolla/redis:2025.1                               \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             redis",
                                "c1265b35218f   registry.osism.tech/kolla/memcached:2025.1                           \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes (healthy)             memcached",
                                "377523689e65   registry.osism.tech/kolla/cron:2025.1                                \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       cron",
                                "84642d628a7d   registry.osism.tech/kolla/kolla-toolbox:2025.1                       \"dumb-init --single-\u2026\"   29 minutes ago   Up 29 minutes                       kolla_toolbox",
                                "7398de513872   registry.osism.tech/kolla/fluentd:2025.1                             \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-2",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      c5b650628899   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2025.1    a7d5f8f5723c   23 hours ago   677MB",
                                "registry.osism.tech/kolla/rabbitmq                            2025.1    aa8dd9f836a1   23 hours ago   345MB",
                                "registry.osism.tech/kolla/fluentd                             2025.1    57639eec9828   23 hours ago   585MB",
                                "registry.osism.tech/kolla/grafana                             2025.1    081a0522ec5e   23 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keepalived                          2025.1    b0edfae0bd90   23 hours ago   282MB",
                                "registry.osism.tech/kolla/haproxy                             2025.1    310d29bed7e8   23 hours ago   280MB",
                                "registry.osism.tech/kolla/memcached                           2025.1    a4e5aa154c9e   23 hours ago   272MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2025.1    72dff88a0a21   23 hours ago   1.54GB",
                                "registry.osism.tech/kolla/cron                                2025.1    69f3c33abd1e   23 hours ago   271MB",
                                "registry.osism.tech/kolla/opensearch                          2025.1    3519bf3e7599   23 hours ago   1.56GB",
                                "registry.osism.tech/kolla/proxysql                            2025.1    29c0cba268b0   23 hours ago   422MB",
                                "registry.osism.tech/kolla/redis                               2025.1    64a84a1dc202   23 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2025.1    717b0336da28   23 hours ago   278MB",
                                "registry.osism.tech/kolla/mariadb-server                      2025.1    836508045bfe   23 hours ago   458MB",
                                "registry.osism.tech/kolla/horizon                             2025.1    98a77f25bcec   23 hours ago   1.2GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2025.1    791ade9ee1bc   23 hours ago   288MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2025.1    534b26335c30   23 hours ago   288MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2025.1    77cad0b71d1f   23 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2025.1    ffdbf401c588   23 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2025.1    309be35cfafe   23 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2025.1    98f129deee0e   23 hours ago   363MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2025.1    ce49e82a896f   23 hours ago   311MB",
                                "registry.osism.tech/kolla/designate-central                   2025.1    92e1c2528bd9   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2025.1    cb3294db7597   23 hours ago   1GB",
                                "registry.osism.tech/kolla/designate-api                       2025.1    c9d9ba583526   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-mdns                      2025.1    6260ddf21d62   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-producer                  2025.1    0fc627b741c8   23 hours ago   999MB",
                                "registry.osism.tech/kolla/designate-worker                    2025.1    7c1af0100cae   23 hours ago   1GB",
                                "registry.osism.tech/kolla/magnum-api                          2025.1    1fe38dcc70df   23 hours ago   1.15GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2025.1    e82724c4bb3d   23 hours ago   1.26GB",
                                "registry.osism.tech/kolla/cinder-volume                       2025.1    1ef7a53c062c   23 hours ago   1.79GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2025.1    53d46e1bbd9d   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-api                          2025.1    40941f417dfe   23 hours ago   1.43GB",
                                "registry.osism.tech/kolla/cinder-backup                       2025.1    033daf13a62d   23 hours ago   1.44GB",
                                "registry.osism.tech/kolla/neutron-server                      2025.1    33584a908e3c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2025.1    203d026e905a   23 hours ago   999MB",
                                "registry.osism.tech/kolla/barbican-api                        2025.1    29406fa4a09f   23 hours ago   998MB",
                                "registry.osism.tech/kolla/barbican-worker                     2025.1    3bb2c4711a36   23 hours ago   999MB",
                                "registry.osism.tech/kolla/keystone-ssh                        2025.1    d2956dec249b   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2025.1    93c2583c8be6   23 hours ago   1.08GB",
                                "registry.osism.tech/kolla/keystone-fernet                     2025.1    8505216f1a5d   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-api                         2025.1    f85caa5bac05   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2025.1    372f8c78f993   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-worker                      2025.1    7a6f2ffd3ac0   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2025.1    edf884456dc4   23 hours ago   1.07GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2025.1    7e4c16624843   23 hours ago   1.05GB",
                                "registry.osism.tech/kolla/nova-api                            2025.1    c643537b5c4c   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-conductor                      2025.1    da0011ad07d6   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2025.1    0984cc5e3ca5   23 hours ago   1.23GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2025.1    ffe8abdb9f27   23 hours ago   1.39GB",
                                "registry.osism.tech/kolla/placement-api                       2025.1    37839dd09fa4   23 hours ago   990MB",
                                "registry.osism.tech/kolla/glance-api                          2025.1    278fe5153fb7   23 hours ago   1.12GB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2025.1    3848e9fd6348   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-northd                          2025.1    90d74abfafe0   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2025.1    e8b493f28635   23 hours ago   295MB",
                                "registry.osism.tech/kolla/ovn-controller                      2025.1    44955f858e65   23 hours ago   296MB",
                                "registry.osism.tech/kolla/ovn-sb-db-relay                     2025.1    fa5efc395605   23 hours ago   295MB",
                                "+ sh -c /opt/configuration/scripts/check-services.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# Ceph status",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# Ceph status'",
                                "+ echo",
                                "+ ceph -s",
                                "  cluster:",
                                "    id:     11111111-1111-1111-1111-111111111111",
                                "    health: HEALTH_OK",
                                " ",
                                "  services:",
                                "    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 27m)",
                                "    mgr: testbed-node-0(active, since 16m), standbys: testbed-node-2, testbed-node-1",
                                "    mds: 1/1 daemons up, 2 standby",
                                "    osd: 6 osds: 6 up (since 24m), 6 in (since 24m)",
                                "    rgw: 3 daemons active (3 hosts, 1 zones)",
                                " ",
                                "  data:",
                                "    volumes: 1/1 healthy",
                                "    pools:   14 pools, 401 pgs",
                                "    objects: 556 objects, 2.2 GiB",
                                "    usage:   7.1 GiB used, 113 GiB / 120 GiB avail",
                                "    pgs:     401 active+clean",
                                " ",
                                "",
                                "# Ceph versions",
                                "",
                                "+ echo",
                                "+ echo '# Ceph versions'",
                                "+ echo",
                                "+ ceph versions",
                                "{",
                                "    \"mon\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"mgr\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"osd\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6",
                                "    },",
                                "    \"mds\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"rgw\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"overall\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18",
                                "    }",
                                "}",
                                "",
                                "# Ceph OSD tree",
                                "",
                                "+ echo",
                                "+ echo '# Ceph OSD tree'",
                                "+ echo",
                                "+ ceph osd df tree",
                                "ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA      OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              ",
                                "-1         0.11691         -  120 GiB  7.1 GiB   6.7 GiB    6 KiB  430 MiB  113 GiB  5.92  1.00    -          root default           ",
                                "-3         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-3",
                                " 1    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.2 GiB    1 KiB   70 MiB   19 GiB  6.45  1.09  204      up          osd.1          ",
                                " 4    hdd  0.01949   1.00000   20 GiB  1.1 GiB   1.0 GiB    1 KiB   74 MiB   19 GiB  5.38  0.91  186      up          osd.4          ",
                                "-7         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-4",
                                " 0    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.2 GiB    1 KiB   74 MiB   19 GiB  6.59  1.11  175      up          osd.0          ",
                                " 3    hdd  0.01949   1.00000   20 GiB  1.0 GiB  1003 MiB    1 KiB   70 MiB   19 GiB  5.24  0.89  213      up          osd.3          ",
                                "-5         0.03897         -   40 GiB  2.4 GiB   2.2 GiB    2 KiB  143 MiB   38 GiB  5.92  1.00    -              host testbed-node-5",
                                " 2    hdd  0.01949   1.00000   20 GiB  1.3 GiB   1.3 GiB    1 KiB   70 MiB   19 GiB  6.69  1.13  206      up          osd.2          ",
                                " 5    hdd  0.01949   1.00000   20 GiB  1.0 GiB   979 MiB    1 KiB   74 MiB   19 GiB  5.14  0.87  186      up          osd.5          ",
                                "                       TOTAL  120 GiB  7.1 GiB   6.7 GiB  9.3 KiB  430 MiB  113 GiB  5.92                                            ",
                                "MIN/MAX VAR: 0.87/1.13  STDDEV: 0.67",
                                "",
                                "# Ceph monitor status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph monitor status'",
                                "+ echo",
                                "+ ceph mon stat",
                                "e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2",
                                "",
                                "# Ceph quorum status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph quorum status'",
                                "+ echo",
                                "+ ceph quorum_status",
                                "+ jq",
                                "{",
                                "  \"election_epoch\": 8,",
                                "  \"quorum\": [",
                                "    0,",
                                "    1,",
                                "    2",
                                "  ],",
                                "  \"quorum_names\": [",
                                "    \"testbed-node-0\",",
                                "    \"testbed-node-1\",",
                                "    \"testbed-node-2\"",
                                "  ],",
                                "  \"quorum_leader_name\": \"testbed-node-0\",",
                                "  \"quorum_age\": 1680,",
                                "  \"features\": {",
                                "    \"quorum_con\": \"4540138322906710015\",",
                                "    \"quorum_mon\": [",
                                "      \"kraken\",",
                                "      \"luminous\",",
                                "      \"mimic\",",
                                "      \"osdmap-prune\",",
                                "      \"nautilus\",",
                                "      \"octopus\",",
                                "      \"pacific\",",
                                "      \"elector-pinging\",",
                                "      \"quincy\",",
                                "      \"reef\"",
                                "    ]",
                                "  },",
                                "  \"monmap\": {",
                                "    \"epoch\": 1,",
                                "    \"fsid\": \"11111111-1111-1111-1111-111111111111\",",
                                "    \"modified\": \"2026-03-13T00:51:25.351727Z\",",
                                "    \"created\": \"2026-03-13T00:51:25.351727Z\",",
                                "    \"min_mon_release\": 18,",
                                "    \"min_mon_release_name\": \"reef\",",
                                "    \"election_strategy\": 1,",
                                "    \"disallowed_leaders: \": \"\",",
                                "    \"stretch_mode\": false,",
                                "    \"tiebreaker_mon\": \"\",",
                                "    \"removed_ranks: \": \"\",",
                                "    \"features\": {",
                                "      \"persistent\": [",
                                "        \"kraken\",",
                                "        \"luminous\",",
                                "        \"mimic\",",
                                "        \"osdmap-prune\",",
                                "        \"nautilus\",",
                                "        \"octopus\",",
                                "        \"pacific\",",
                                "        \"elector-pinging\",",
                                "        \"quincy\",",
                                "        \"reef\"",
                                "      ],",
                                "      \"optional\": []",
                                "    },",
                                "    \"mons\": [",
                                "      {",
                                "        \"rank\": 0,",
                                "        \"name\": \"testbed-node-0\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.10:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.10:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.10:6789/0\",",
                                "        \"public_addr\": \"192.168.16.10:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 1,",
                                "        \"name\": \"testbed-node-1\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.11:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.11:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.11:6789/0\",",
                                "        \"public_addr\": \"192.168.16.11:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 2,",
                                "        \"name\": \"testbed-node-2\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.12:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.12:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.12:6789/0\",",
                                "        \"public_addr\": \"192.168.16.12:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      }",
                                "    ]",
                                "  }",
                                "}",
                                "",
                                "# Ceph free space status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph free space status'",
                                "+ echo",
                                "+ ceph df",
                                "--- RAW STORAGE ---",
                                "CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED",
                                "hdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92",
                                "TOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.92",
                                " ",
                                "--- POOLS ---",
                                "POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL",
                                ".mgr                        1    1  577 KiB        2  1.1 MiB      0     53 GiB",
                                "cephfs_data                 2   32      0 B        0      0 B      0     35 GiB",
                                "cephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB",
                                "default.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.control         6   32      0 B        8      0 B      0     35 GiB",
                                "default.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB",
                                "default.rgw.meta            8   32      0 B        0      0 B      0     35 GiB",
                                ".rgw.root                   9   32  3.9 KiB        8   64 KiB      0     53 GiB",
                                "backups                    10   32     19 B        2   12 KiB      0     35 GiB",
                                "volumes                    11   32     19 B        2   12 KiB      0     35 GiB",
                                "images                     12   32  2.2 GiB      299  6.7 GiB   5.92     35 GiB",
                                "metrics                    13   32     19 B        2   12 KiB      0     35 GiB",
                                "vms                        14   32     19 B        2   12 KiB      0     35 GiB",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ ! -e /etc/redhat-release ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6ab5989f-2cab-410b-9a43-0ddd7b56424d (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 01:19:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 6ab5989f-2cab-410b-9a43-0ddd7b56424d (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Friday 13 March 2026  01:20:00 +0000 (0:00:00.228)       0:00:00.228 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Friday 13 March 2026  01:20:01 +0000 (0:00:01.326)       0:00:01.554 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Friday 13 March 2026  01:20:02 +0000 (0:00:01.106)       0:00:02.661 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Friday 13 March 2026  01:20:07 +0000 (0:00:04.184)       0:00:06.846 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:20:07 +0000 (0:00:00.473)       0:00:07.319 ********** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.18s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.33s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.11s",
                                "Gather facts for all hosts ---------------------------------------------- 0.47s",
                                "+ osism validate ceph-mons",
                                "",
                                "PLAY [Ceph validate mons] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Friday 13 March 2026  01:20:25 +0000 (0:00:00.434)       0:00:00.434 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Friday 13 March 2026  01:20:25 +0000 (0:00:00.810)       0:00:01.245 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Friday 13 March 2026  01:20:26 +0000 (0:00:00.916)       0:00:02.161 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Friday 13 March 2026  01:20:26 +0000 (0:00:00.133)       0:00:02.294 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Friday 13 March 2026  01:20:27 +0000 (0:00:00.293)       0:00:02.588 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Friday 13 March 2026  01:20:28 +0000 (0:00:01.219)       0:00:03.808 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Friday 13 March 2026  01:20:28 +0000 (0:00:00.285)       0:00:04.093 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:20:29 +0000 (0:00:00.471)       0:00:04.565 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mon is not running] ********************",
                                "Friday 13 March 2026  01:20:29 +0000 (0:00:00.301)       0:00:04.867 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mon is running] ************************",
                                "Friday 13 March 2026  01:20:29 +0000 (0:00:00.298)       0:00:05.165 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:20:30 +0000 (0:00:00.442)       0:00:05.607 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:20:30 +0000 (0:00:00.275)       0:00:05.883 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:20:30 +0000 (0:00:00.255)       0:00:06.139 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:30 +0000 (0:00:00.264)       0:00:06.404 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.069)       0:00:06.473 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.067)       0:00:06.540 ********** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.088)       0:00:06.629 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.232)       0:00:06.862 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare quorum test vars] ************************************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.234)       0:00:07.097 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get monmap info from one mon container] **********************************",
                                "Friday 13 March 2026  01:20:31 +0000 (0:00:00.122)       0:00:07.220 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set quorum test data] ****************************************************",
                                "Friday 13 March 2026  01:20:33 +0000 (0:00:01.759)       0:00:08.979 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail quorum test if not all monitors are in quorum] **********************",
                                "Friday 13 March 2026  01:20:34 +0000 (0:00:00.476)       0:00:09.456 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass quorum test if all monitors are in quorum] **************************",
                                "Friday 13 March 2026  01:20:34 +0000 (0:00:00.120)       0:00:09.577 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set fsid test vars] ******************************************************",
                                "Friday 13 March 2026  01:20:34 +0000 (0:00:00.335)       0:00:09.913 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail Cluster FSID test if FSID does not match configuration] *************",
                                "Friday 13 March 2026  01:20:34 +0000 (0:00:00.303)       0:00:10.217 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass Cluster FSID test if it matches configuration] **********************",
                                "Friday 13 March 2026  01:20:34 +0000 (0:00:00.110)       0:00:10.327 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare status test vars] ************************************************",
                                "Friday 13 March 2026  01:20:35 +0000 (0:00:00.123)       0:00:10.451 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather status data] ******************************************************",
                                "Friday 13 March 2026  01:20:35 +0000 (0:00:00.126)       0:00:10.578 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set health test data] ****************************************************",
                                "Friday 13 March 2026  01:20:36 +0000 (0:00:01.516)       0:00:12.094 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable] *************************",
                                "Friday 13 March 2026  01:20:36 +0000 (0:00:00.300)       0:00:12.395 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if health is acceptable] *****************************",
                                "Friday 13 March 2026  01:20:37 +0000 (0:00:00.125)       0:00:12.520 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable (strict)] ****************",
                                "Friday 13 March 2026  01:20:37 +0000 (0:00:00.144)       0:00:12.664 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if status is OK (strict)] ****************************",
                                "Friday 13 March 2026  01:20:37 +0000 (0:00:00.288)       0:00:12.953 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Friday 13 March 2026  01:20:37 +0000 (0:00:00.134)       0:00:13.088 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Friday 13 March 2026  01:20:37 +0000 (0:00:00.263)       0:00:13.351 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:20:38 +0000 (0:00:00.254)       0:00:13.606 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:20:39 +0000 (0:00:01.696)       0:00:15.302 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:20:40 +0000 (0:00:00.270)       0:00:15.572 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:40 +0000 (0:00:00.267)       0:00:15.839 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:40 +0000 (0:00:00.071)       0:00:15.911 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:20:40 +0000 (0:00:00.073)       0:00:15.984 ********** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Friday 13 March 2026  01:20:40 +0000 (0:00:00.075)       0:00:16.060 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:20:42 +0000 (0:00:01.459)       0:00:17.520 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-03-13T01:20:25+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:20:42 +0000 (0:00:00.785)       0:00:18.305 ********** ",
                                "=============================================================================== ",
                                "Get monmap info from one mon container ---------------------------------- 1.76s",
                                "Aggregate test results step one ----------------------------------------- 1.70s",
                                "Gather status data ------------------------------------------------------ 1.52s",
                                "Write report file ------------------------------------------------------- 1.46s",
                                "Get container info ------------------------------------------------------ 1.22s",
                                "Create report output directory ------------------------------------------ 0.92s",
                                "Get timestamp for report file ------------------------------------------- 0.81s",
                                "Print report file information ------------------------------------------- 0.79s",
                                "Set quorum test data ---------------------------------------------------- 0.48s",
                                "Set test result to passed if container is existing ---------------------- 0.47s",
                                "Set test result to passed if ceph-mon is running ------------------------ 0.44s",
                                "Pass quorum test if all monitors are in quorum -------------------------- 0.34s",
                                "Set fsid test vars ------------------------------------------------------ 0.30s",
                                "Prepare test data ------------------------------------------------------- 0.30s",
                                "Set health test data ---------------------------------------------------- 0.30s",
                                "Set test result to failed if ceph-mon is not running -------------------- 0.30s",
                                "Prepare test data for container existance test -------------------------- 0.29s",
                                "Fail cluster-health if health is not acceptable (strict) ---------------- 0.29s",
                                "Set test result to failed if container is missing ----------------------- 0.29s",
                                "Aggregate test results step one ----------------------------------------- 0.28s",
                                "+ osism validate ceph-mgrs",
                                "",
                                "PLAY [Ceph validate mgrs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Friday 13 March 2026  01:21:00 +0000 (0:00:00.448)       0:00:00.448 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Friday 13 March 2026  01:21:01 +0000 (0:00:00.816)       0:00:01.265 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Friday 13 March 2026  01:21:02 +0000 (0:00:01.005)       0:00:02.271 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Friday 13 March 2026  01:21:02 +0000 (0:00:00.129)       0:00:02.400 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Friday 13 March 2026  01:21:02 +0000 (0:00:00.281)       0:00:02.681 ********** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Friday 13 March 2026  01:21:03 +0000 (0:00:00.974)       0:00:03.656 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Friday 13 March 2026  01:21:03 +0000 (0:00:00.278)       0:00:03.934 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:04 +0000 (0:00:00.495)       0:00:04.430 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mgr is not running] ********************",
                                "Friday 13 March 2026  01:21:04 +0000 (0:00:00.297)       0:00:04.728 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mgr is running] ************************",
                                "Friday 13 March 2026  01:21:04 +0000 (0:00:00.312)       0:00:05.041 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:21:05 +0000 (0:00:00.508)       0:00:05.550 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:21:05 +0000 (0:00:00.252)       0:00:05.802 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:21:05 +0000 (0:00:00.251)       0:00:06.055 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.265)       0:00:06.321 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.069)       0:00:06.390 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.079)       0:00:06.470 ********** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.073)       0:00:06.543 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.244)       0:00:06.788 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Define mgr module test vars] *********************************************",
                                "Friday 13 March 2026  01:21:06 +0000 (0:00:00.246)       0:00:07.034 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather list of mgr modules] **********************************************",
                                "Friday 13 March 2026  01:21:07 +0000 (0:00:00.144)       0:00:07.179 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Parse mgr module list from json] *****************************************",
                                "Friday 13 March 2026  01:21:09 +0000 (0:00:02.060)       0:00:09.240 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Extract list of enabled mgr modules] *************************************",
                                "Friday 13 March 2026  01:21:09 +0000 (0:00:00.429)       0:00:09.670 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail test if mgr modules are disabled that should be enabled] ************",
                                "Friday 13 March 2026  01:21:09 +0000 (0:00:00.303)       0:00:09.974 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass test if required mgr modules are enabled] ***************************",
                                "Friday 13 March 2026  01:21:09 +0000 (0:00:00.139)       0:00:10.113 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Friday 13 March 2026  01:21:10 +0000 (0:00:00.155)       0:00:10.269 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Friday 13 March 2026  01:21:10 +0000 (0:00:00.251)       0:00:10.521 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:21:10 +0000 (0:00:00.248)       0:00:10.769 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:21:11 +0000 (0:00:01.274)       0:00:12.044 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:21:12 +0000 (0:00:00.253)       0:00:12.297 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:12 +0000 (0:00:00.243)       0:00:12.540 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:12 +0000 (0:00:00.069)       0:00:12.609 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:12 +0000 (0:00:00.067)       0:00:12.677 ********** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Friday 13 March 2026  01:21:12 +0000 (0:00:00.239)       0:00:12.917 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:21:14 +0000 (0:00:01.264)       0:00:14.182 ********** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-03-13T01:21:01+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:21:14 +0000 (0:00:00.389)       0:00:14.571 ********** ",
                                "=============================================================================== ",
                                "Gather list of mgr modules ---------------------------------------------- 2.06s",
                                "Aggregate test results step one ----------------------------------------- 1.28s",
                                "Write report file ------------------------------------------------------- 1.26s",
                                "Create report output directory ------------------------------------------ 1.01s",
                                "Get container info ------------------------------------------------------ 0.97s",
                                "Get timestamp for report file ------------------------------------------- 0.82s",
                                "Set test result to passed if ceph-mgr is running ------------------------ 0.51s",
                                "Set test result to passed if container is existing ---------------------- 0.50s",
                                "Parse mgr module list from json ----------------------------------------- 0.43s",
                                "Print report file information ------------------------------------------- 0.39s",
                                "Flush handlers ---------------------------------------------------------- 0.38s",
                                "Set test result to failed if ceph-mgr is not running -------------------- 0.31s",
                                "Extract list of enabled mgr modules ------------------------------------- 0.30s",
                                "Prepare test data ------------------------------------------------------- 0.30s",
                                "Prepare test data for container existance test -------------------------- 0.28s",
                                "Set test result to failed if container is missing ----------------------- 0.28s",
                                "Aggregate test results step three --------------------------------------- 0.27s",
                                "Aggregate test results step two ----------------------------------------- 0.25s",
                                "Aggregate test results step two ----------------------------------------- 0.25s",
                                "Aggregate test results step one ----------------------------------------- 0.25s",
                                "+ osism validate ceph-osds",
                                "",
                                "PLAY [Ceph validate OSDs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Friday 13 March 2026  01:21:26 +0000 (0:00:00.320)       0:00:00.320 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Friday 13 March 2026  01:21:27 +0000 (0:00:00.678)       0:00:00.998 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Friday 13 March 2026  01:21:27 +0000 (0:00:00.382)       0:00:01.380 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Friday 13 March 2026  01:21:28 +0000 (0:00:00.689)       0:00:02.070 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Friday 13 March 2026  01:21:28 +0000 (0:00:00.122)       0:00:02.193 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Friday 13 March 2026  01:21:28 +0000 (0:00:00.125)       0:00:02.318 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Friday 13 March 2026  01:21:28 +0000 (0:00:00.276)       0:00:02.594 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Friday 13 March 2026  01:21:28 +0000 (0:00:00.129)       0:00:02.724 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate total number of OSDs in cluster] *******************************",
                                "Friday 13 March 2026  01:21:29 +0000 (0:00:00.291)       0:00:03.015 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:29 +0000 (0:00:00.606)       0:00:03.621 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers on host] *********************************",
                                "Friday 13 March 2026  01:21:30 +0000 (0:00:00.271)       0:00:03.893 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8c865601b389d1fe688c59bbce7c9c3bc0724c5899d822fc61c74068974e61fe', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b7147f1bef5ea7241362eea7f8538510c31ad51819c12903214405857b7d82ef', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4e1aa8565c9877cdf1f897b28d56ab3187247dd869e3f51bbf01bd512df70508', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '06b1cc71ccdec33e79f0b4eee83c0db488ddfd5ec1efa160310077073e5d18b3', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8e31420c6242de113d89aea4314577257644b2d30d8144daef0da155da570ba4', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '219a66a5a5bf1f0b561aeb7d424fadef452117281bb8152e3b64f5d3088edb78', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '838a854ce875d0fc141e74bf22c0a7158d3114abf1ba759e85a6e59b70c2acb1', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'f3943bcde8de3d5b646e46f400d2545dc9779a0f94c9df66985e046139e9e418', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ba676f3da3ee3f7f55f8d5410d9bc077c2fbee97d9f8442517186928e0e888b3', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a0baa27dd316ac2ef80e1cdc784cf4b694009581bd83040fb77e61232d4fb483', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': '1e74e68e4ff6794127d01d9e1b32b0fc09cbdf0c82b23241e09c044508f7496f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': '47641c347c343cbe4b73bb4d590d02d728fa115e90a0086256ecafeb5e4974de', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e5bbad3d6dce42daa045230503e1ad52f59d3e63073c9b4587768810646c39e0', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'b616fd0ff3e59d7fc1c8f57d1a304fdec40e0b1c4ffde4bd698e4c9f5b5a18c1', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '006940d481062f8b2bf4b1b00067947de48a516d70b5ea5a8f48671a62d95e2c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e6c2bb5ee88ef518b998b52ae3653df452ac14b573c770bb161133d5aa820114', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '80cafc4a6dd432a2858bdd8ea6ff67fca743dd46eaac1016ae21b2b87cef727c', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '3e7ed081145c3d2b7bf8bbfe818f1b6eb0567923e5af61704515ba8d0ae9c02c', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd60cd154363827e30cbe4f4d9b5e012768419f516a1afc13c1ee5ceeb6230253', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'b95bcbea35d52acc6645634e35951aeb7a7439e77bc3e197b9e589de86441fdd', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '8d3afd5541fd1475a55f8b97aef4c69e91426749be733dbd666f7163cf5de904', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'ee4e96ce1c56963a8affa28df4358c450cc7d756df5439da37b5d755c8871943', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6ed0baee43bd02ab2f73e8e6b9e3dfd032583a669d7bf66b95073bf0e946e3a7', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'fe0432ecce244f4ef2fea3a7eb2166018fab34e1c6e328da7cf083287d215a44', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '6e5dcda8d1bcd9cf76d0256ba123cdeaeaf3b228722799666e920c8f1a93953e', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'f989112bf0b79d4606f895117eb247f2750c570247cfa7c3894e7373d20975b0', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '02f04d5ec85a7b0bde6c2e42443537c842c00aa31ecd16ccb5a70b0d90f02a84', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '56351f94e56c5920e8517dfcd0c7a9dc34f50bf8b12aca32e05a7804a6760937', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '3303bf2bba0dd445aa1761d756c623248ccfffacd119372c68568e1e104b029e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '4b1511bac231627cd4efed9cc7a0aacf8d66dde51f2448944f4aeaf81e99ce60', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'be870e67c539dee64c08627ffea04e3ac232da50d0727ac2686ef13de2a225c5', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'a80efbde99dc30df91150ce781bc04ac56c15d1ceb2a9b1bc7cbff7b99ddb4e3', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '05a44d95a05c7995789ae4bcd886fec8de74b046a87c1ceae2bcafe66782be5c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '7cf69b4487569e329087c1913e4338f59af3982ddfe9410f1cf88202b6715926', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '767104052acc36a839debca40d71b9d8c6986272d1bd940553259c82b7f82d2a', 'image': 'registry.osism.tech/kolla/nova-compute:2025.1', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b731b3ceaa9ba7f81628c9432a7cbcdee4b64c17d69f61494ab95e1fd973af63', 'image': 'registry.osism.tech/kolla/nova-libvirt:2025.1', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1651b8e31a45bb8d0b9f8c6e901cbb59fa9e1f6e736f41ed2f966b07b6dcb18d', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '630b5a32fbb3395f252e15063c19a5d4ee9b7864627253773df5cbf23ed00ee6', 'image': 'registry.osism.tech/kolla/nova-ssh:2025.1', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '32eab2929619d7d147289ad2798d6157ff7ff393a74f5df88592c0ac038b95f0', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '1cfaec017d1a28c40a0f888e3e6d7c67849fdc35f0c7fa0093885015957e055b', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2025.1', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'd713c8de9f0dc8f642f1290a61846f1c29c9a79c9ad133fc8b252485630c678b', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2025.1', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '86cd5d46c38aaedf25869ba48fdb9fd1763b11297da05dab58b5c654c2d949bb', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2025.1', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f515baadbd95b8e302c19b43f5701960b2f8a7becb4887fa5dfe39919557bb87', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2025.1', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 14 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c6fbb16e8cf12df44a95ab5c3d432545a320eba8e2f29507a1be7fa6dd8a64ba', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 21 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b4abc0ef2dcb6e787bd375cba32312d3e0ee01099f3a3237f4351bcbd21aa5fb', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8ddb33a2427e56dc1d44c1cf60ba5f09d310b217e628113d00601c2d73fdeb97', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '5fe0fdcc2774f76391ead83aa2df8b5a36fa8a4bf962924dd530d7c6709bd25e', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': 'efebf37ff7bf9063c1f01b72f82e34857c0babb9b9d3d0bd32310d05201ae34f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 24 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fbdf31c96080ceb795826725def5e8a1b0444e30a00fe2f706b72cdb45a214ee', 'image': 'registry.osism.tech/kolla/ovn-controller:2025.1', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'd57416ad822d57921d5d3e0ca2ad56327699af161bc334e768a6bab3ab054285', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2025.1', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 28 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '77597e65fb914d92c76b6c3fed3d7353f14955be3003d22a228780ca5fece769', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2025.1', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 29 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '52659bd4d1973760b326a0b719d38f64ec2e4fb717c5699d0da21ff4b41eddd2', 'image': 'registry.osism.tech/kolla/cron:2025.1', 'name': '/cron', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8fe38e8a78cf46bd7fd5e5d237c9de16213ea5902685c5eadcaa5c100a7fd417', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2025.1', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'f45c4e6b4aff1e7f32694e5aed9c8b49a55076313bbc23c730b8a38a4b7a9239', 'image': 'registry.osism.tech/kolla/fluentd:2025.1', 'name': '/fluentd', 'state': 'running', 'status': 'Up 30 minutes'}) \u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers on host] ********************************",
                                "Friday 13 March 2026  01:21:30 +0000 (0:00:00.408)       0:00:04.301 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed when count of containers is wrong] *************",
                                "Friday 13 March 2026  01:21:30 +0000 (0:00:00.251)       0:00:04.553 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if count matches] ******************************",
                                "Friday 13 March 2026  01:21:31 +0000 (0:00:00.398)       0:00:04.952 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:31 +0000 (0:00:00.277)       0:00:05.230 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers that are not running] ********************",
                                "Friday 13 March 2026  01:21:31 +0000 (0:00:00.247)       0:00:05.477 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers that are not running] *******************",
                                "Friday 13 March 2026  01:21:32 +0000 (0:00:00.297)       0:00:05.774 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Friday 13 March 2026  01:21:32 +0000 (0:00:00.389)       0:00:06.163 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Friday 13 March 2026  01:21:32 +0000 (0:00:00.262)       0:00:06.426 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if all containers are running] *****************",
                                "Friday 13 March 2026  01:21:32 +0000 (0:00:00.242)       0:00:06.668 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:21:33 +0000 (0:00:00.258)       0:00:06.927 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:21:33 +0000 (0:00:00.501)       0:00:07.429 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:21:33 +0000 (0:00:00.219)       0:00:07.648 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.231)       0:00:07.880 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.065)       0:00:07.945 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.072)       0:00:08.017 ********** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.064)       0:00:08.081 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail early due to containers not running] ********************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.234)       0:00:08.316 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:34 +0000 (0:00:00.264)       0:00:08.581 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set _mon_hostname fact] **************************************************",
                                "Friday 13 March 2026  01:21:35 +0000 (0:00:00.284)       0:00:08.865 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get ceph osd tree] *******************************************************",
                                "Friday 13 March 2026  01:21:35 +0000 (0:00:00.682)       0:00:09.547 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Parse osd tree from JSON] ************************************************",
                                "Friday 13 March 2026  01:21:37 +0000 (0:00:01.861)       0:00:11.409 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get OSDs that are not up or in] ******************************************",
                                "Friday 13 March 2026  01:21:37 +0000 (0:00:00.122)       0:00:11.531 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail test if OSDs are not up or in] **************************************",
                                "Friday 13 March 2026  01:21:38 +0000 (0:00:00.309)       0:00:11.840 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Pass test if OSDs are all up and in] *************************************",
                                "Friday 13 March 2026  01:21:38 +0000 (0:00:00.125)       0:00:11.966 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:38 +0000 (0:00:00.150)       0:00:12.117 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [List ceph LVM volumes and collect data] **********************************",
                                "Friday 13 March 2026  01:21:38 +0000 (0:00:00.274)       0:00:12.391 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Parse LVM data as JSON] **************************************************",
                                "Friday 13 March 2026  01:21:40 +0000 (0:00:02.263)       0:00:14.655 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get unencrypted and encrypted OSDs] **************************************",
                                "Friday 13 March 2026  01:21:41 +0000 (0:00:00.488)       0:00:15.143 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of encrypted OSDs does not match] **************************",
                                "Friday 13 March 2026  01:21:41 +0000 (0:00:00.494)       0:00:15.638 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of encrypted OSDs equals count of OSDs] ********************",
                                "Friday 13 March 2026  01:21:42 +0000 (0:00:00.287)       0:00:15.926 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of unencrypted OSDs does not match] ************************",
                                "Friday 13 March 2026  01:21:42 +0000 (0:00:00.450)       0:00:16.376 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************",
                                "Friday 13 March 2026  01:21:42 +0000 (0:00:00.294)       0:00:16.671 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Friday 13 March 2026  01:21:43 +0000 (0:00:00.292)       0:00:16.963 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get CRUSH node data of each OSD host and root node childs] ***************",
                                "Friday 13 March 2026  01:21:43 +0000 (0:00:00.469)       0:00:17.433 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate sub test expression results] ***********************************",
                                "Friday 13 March 2026  01:21:44 +0000 (0:00:00.712)       0:00:18.146 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail test if any sub test failed] ****************************************",
                                "Friday 13 March 2026  01:21:44 +0000 (0:00:00.311)       0:00:18.457 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass test if no sub test failed] *****************************************",
                                "Friday 13 March 2026  01:21:45 +0000 (0:00:00.328)       0:00:18.786 ********** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Friday 13 March 2026  01:21:45 +0000 (0:00:00.324)       0:00:19.110 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Friday 13 March 2026  01:21:46 +0000 (0:00:00.664)       0:00:19.775 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Friday 13 March 2026  01:21:46 +0000 (0:00:00.235)       0:00:20.010 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Friday 13 March 2026  01:21:47 +0000 (0:00:01.511)       0:00:21.522 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Friday 13 March 2026  01:21:47 +0000 (0:00:00.248)       0:00:21.770 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:48 +0000 (0:00:00.272)       0:00:22.043 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:48 +0000 (0:00:00.081)       0:00:22.126 ********** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Friday 13 March 2026  01:21:48 +0000 (0:00:00.067)       0:00:22.194 ********** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Friday 13 March 2026  01:21:48 +0000 (0:00:00.069)       0:00:22.263 ********** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Friday 13 March 2026  01:21:49 +0000 (0:00:01.269)       0:00:23.533 ********** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-03-13T01:21:27+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:21:50 +0000 (0:00:00.576)       0:00:24.109 ********** ",
                                "=============================================================================== ",
                                "List ceph LVM volumes and collect data ---------------------------------- 2.26s",
                                "Get ceph osd tree ------------------------------------------------------- 1.86s",
                                "Aggregate test results step one ----------------------------------------- 1.51s",
                                "Write report file ------------------------------------------------------- 1.27s",
                                "Get CRUSH node data of each OSD host and root node childs --------------- 0.71s",
                                "Create report output directory ------------------------------------------ 0.69s",
                                "Set _mon_hostname fact -------------------------------------------------- 0.68s",
                                "Get timestamp for report file ------------------------------------------- 0.68s",
                                "Set validation result to passed if no test failed ----------------------- 0.66s",
                                "Calculate total number of OSDs in cluster ------------------------------- 0.61s",
                                "Print report file information ------------------------------------------- 0.58s",
                                "Aggregate test results step one ----------------------------------------- 0.50s",
                                "Get unencrypted and encrypted OSDs -------------------------------------- 0.49s",
                                "Parse LVM data as JSON -------------------------------------------------- 0.49s",
                                "Prepare test data ------------------------------------------------------- 0.47s",
                                "Pass if count of encrypted OSDs equals count of OSDs -------------------- 0.45s",
                                "Get list of ceph-osd containers on host --------------------------------- 0.41s",
                                "Set test result to failed when count of containers is wrong ------------- 0.40s",
                                "Get count of ceph-osd containers that are not running ------------------- 0.39s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.38s",
                                "+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ source /etc/os-release",
                                "++ PRETTY_NAME='Ubuntu 24.04.4 LTS'",
                                "++ NAME=Ubuntu",
                                "++ VERSION_ID=24.04",
                                "++ VERSION='24.04.4 LTS (Noble Numbat)'",
                                "++ VERSION_CODENAME=noble",
                                "++ ID=ubuntu",
                                "++ ID_LIKE=debian",
                                "++ HOME_URL=https://www.ubuntu.com/",
                                "++ SUPPORT_URL=https://help.ubuntu.com/",
                                "++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/",
                                "++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy",
                                "++ UBUNTU_CODENAME=noble",
                                "++ LOGO=ubuntu-logo",
                                "+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]",
                                "+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'",
                                "+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "",
                                "# Status of Elasticsearch",
                                "",
                                "+ pushd /opt/configuration/contrib",
                                "+ echo",
                                "+ echo '# Status of Elasticsearch'",
                                "+ echo",
                                "+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s",
                                "OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0",
                                "",
                                "# Status of MariaDB",
                                "+ echo",
                                "+ echo '# Status of MariaDB'",
                                "+ echo",
                                "",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status database",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: a2f1f1b1-1e77-11f1-9a51-b241cc3da751\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.11:3306,192.168.16.12:3306,192.168.16.10:3306\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: d0a8827e-1e77-11f1-a897-db3dd2836140\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.07%\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.00423789\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4818 local commits, 7019 replicated, 92 received\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 23 minutes, 15 seconds\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 149 connected, 1 running\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 209806 total, 0 slow\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 132\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Prometheus",
                                "",
                                "+ echo",
                                "+ echo '# Status of Prometheus'",
                                "+ echo",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy",
                                "Unauthorized",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready",
                                "Unauthorized",
                                "",
                                "# Status of RabbitMQ",
                                "",
                                "+ echo",
                                "+ echo '# Status of RabbitMQ'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status messaging",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 4.1.8\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 27.3.4.1\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 212, Channels: 211, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 221 total, 221 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 111/1024\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 0/0\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 4.1.8\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 27.3.4.1\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 212, Channels: 211, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 221 total, 221 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.6 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 111/1024\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 0/0\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 4.1.8\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 27.3.4.1\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 212, Channels: 211, Queues: 173\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 221 total, 221 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 5.8/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.15 GB (limit: 18.80 GB)\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 111/1024\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 0/0\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Redis",
                                "",
                                "+ echo",
                                "+ echo '# Status of Redis'",
                                "+ echo",
                                "+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j",
                                "TCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.001782s;;;0.000000;10.000000",
                                "",
                                "# Create backup of MariaDB database",
                                "",
                                "+ popd",
                                "+ echo",
                                "+ echo '# Create backup of MariaDB database'",
                                "+ echo",
                                "+ osism apply mariadb_backup -e mariadb_backup_type=full",
                                "\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f9efa469-6931-4d78-a59c-d3d2ae5d6be5 (mariadb_backup) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 01:22:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task f9efa469-6931-4d78-a59c-d3d2ae5d6be5 (mariadb_backup) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Friday 13 March 2026  01:22:38 +0000 (0:00:00.168)       0:00:00.168 ********** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Friday 13 March 2026  01:22:38 +0000 (0:00:00.329)       0:00:00.498 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Friday 13 March 2026  01:22:39 +0000 (0:00:00.569)       0:00:01.068 ********** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Friday 13 March 2026  01:22:39 +0000 (0:00:00.424)       0:00:01.493 ********** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB container facts] ***********************************",
                                "Friday 13 March 2026  01:22:40 +0000 (0:00:00.532)       0:00:02.025 ********** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Taking full database backup via Mariabackup] *******************",
                                "Friday 13 March 2026  01:22:43 +0000 (0:00:03.318)       0:00:05.344 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Friday 13 March 2026  01:22:59 +0000 (0:00:16.143)       0:00:21.487 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Friday 13 March 2026  01:23:00 +0000 (0:00:00.294)       0:00:21.782 ********** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:23:00 +0000 (0:00:00.381)       0:00:22.164 ********** ",
                                "=============================================================================== ",
                                "mariadb : Taking full database backup via Mariabackup ------------------ 16.14s",
                                "mariadb : Get MariaDB container facts ----------------------------------- 3.32s",
                                "Group hosts based on enabled services ----------------------------------- 0.57s",
                                "mariadb : include_tasks ------------------------------------------------- 0.53s",
                                "mariadb : Group MariaDB hosts based on shards --------------------------- 0.42s",
                                "Include mariadb post-upgrade.yml ---------------------------------------- 0.38s",
                                "Group hosts based on Kolla action --------------------------------------- 0.33s",
                                "Include mariadb post-deploy.yml ----------------------------------------- 0.29s",
                                "+ sh -c /opt/configuration/scripts/check/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ export OS_CLOUD=admin",
                                "+ OS_CLOUD=admin",
                                "+ echo",
                                "",
                                "# OpenStack endpoints",
                                "",
                                "+ echo '# OpenStack endpoints'",
                                "+ echo",
                                "+ openstack endpoint list",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| 1caec2687c60456cbc7822a383782bc2 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |",
                                "| 201a07ec02d14ca09fd330558a3dd6fe | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |",
                                "| 20590df4ee8a4f86a893a053d0886293 | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |",
                                "| 2140ae4e39674fab810e3d189fabebff | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |",
                                "| 37fead48967943f8aa6a4d759200a29e | RegionOne | cinder       | block-storage   | True    | public    | https://api.testbed.osism.xyz:8776/v3                               |",
                                "| 394cc09f89b043938d746ffb14db3f0d | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |",
                                "| 3ec39bdb982742749096453c50e1a2dc | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |",
                                "| 51ae298bdeab4cfaab67e021b474bb2b | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |",
                                "| 571ffbcfe3724f59972ee4b86febf04c | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |",
                                "| 7804c9d6b7384114bbd1ab1424c980f8 | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |",
                                "| 7e356053703344fabf9de43c4a88ad33 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |",
                                "| 81bcdaf0ddda4ae993189e130798c6e5 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |",
                                "| 985d48b634c2474aae6827f3508edab8 | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |",
                                "| 9bd56d8861d74b6690965436ba896a33 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |",
                                "| a29638ae2912484cae5a2933c7f96f95 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |",
                                "| b498e434c068462f903caa8e515293f5 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |",
                                "| b96bff7722bc4d6ea39ed79e83829c57 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |",
                                "| d20fbe52d0424575b1f01a1ef481fbc6 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |",
                                "| d64f2d9293dc4fbb896e915c9e23aca7 | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |",
                                "| dcc95226d4eb4c32a28c7c55ead33887 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |",
                                "| e415e18e83734c209dd66ba9d74eb6be | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |",
                                "| e94e8394ec4e4a7a98101bfca6c4ee20 | RegionOne | cinder       | block-storage   | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3                           |",
                                "| f5d9ad80054d4ff7a86d8175b186ef9b | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |",
                                "| fa7634d91e82403c87361c99d3d62137 | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "",
                                "# Cinder",
                                "",
                                "+ echo",
                                "+ echo '# Cinder'",
                                "+ echo",
                                "+ openstack volume service list",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| Binary           | Host                       | Zone     | Status  | State | Updated At                 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-03-13T01:22:58.000000 |",
                                "| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-03-13T01:22:59.000000 |",
                                "| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-03-13T01:22:58.000000 |",
                                "| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:22:59.000000 |",
                                "| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:23:04.000000 |",
                                "| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-03-13T01:23:05.000000 |",
                                "| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-03-13T01:23:03.000000 |",
                                "| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-03-13T01:23:05.000000 |",
                                "| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-03-13T01:23:06.000000 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "",
                                "# Neutron",
                                "",
                                "+ echo",
                                "+ echo '# Neutron'",
                                "+ echo",
                                "+ openstack network agent list",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |",
                                "| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "+ openstack network service provider list",
                                "+---------------+------+---------+",
                                "| Service Type  | Name | Default |",
                                "+---------------+------+---------+",
                                "| L3_ROUTER_NAT | ovn  | True    |",
                                "+---------------+------+---------+",
                                "",
                                "# Nova",
                                "",
                                "+ echo",
                                "+ echo '# Nova'",
                                "+ echo",
                                "+ openstack compute service list",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| 34696851-f8b5-44c1-abf2-1235a653e4ab | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |",
                                "| f87f6b66-58f0-453b-b277-6f795b5e975b | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |",
                                "| 408149be-f79f-42b0-b950-5da3ddcb592b | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-03-13T01:23:09.000000 |",
                                "| 2b210f2d-6743-4ed5-9797-fe8a1fe9dda5 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-03-13T01:23:07.000000 |",
                                "| e28b9961-0e3a-441a-b16e-c2e70cd3432c | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-03-13T01:23:08.000000 |",
                                "| b7b61d6d-742b-4109-b14d-dcbe0ca96e1c | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-03-13T01:23:08.000000 |",
                                "| 39372ed5-73ea-46d1-bec2-b9d98d410998 | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-03-13T01:23:11.000000 |",
                                "| babb32bb-811a-4abc-857e-666a13237230 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-03-13T01:23:11.000000 |",
                                "| 7f4b5cac-2b39-4e9e-9df5-5fd5e72908b0 | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-03-13T01:23:12.000000 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "+ openstack hypervisor list",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| 36ff5121-2fb8-42f3-8f4f-36344af1223b | testbed-node-3      | QEMU            | 192.168.16.13 | up    |",
                                "| b4594c09-21eb-4c9a-9400-487c30120c44 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |",
                                "| 0dccbc3a-9367-460d-ae63-5bb2a224353c | testbed-node-5      | QEMU            | 192.168.16.15 | up    |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "+ echo",
                                "",
                                "# Run OpenStack test play",
                                "",
                                "+ echo '# Run OpenStack test play'",
                                "+ echo",
                                "+ osism apply --environment openstack test",
                                "\u001b[32m2026-03-13 01:23:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m",
                                "\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m",
                                "\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask cff00ddd-7547-4db0-9f4c-3644ccf273b4 (test) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 01:23:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task cff00ddd-7547-4db0-9f4c-3644ccf273b4 (test) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Create test project] *****************************************************",
                                "",
                                "TASK [Create test domain] ******************************************************",
                                "Friday 13 March 2026  01:23:32 +0000 (0:00:00.051)       0:00:00.051 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test-admin user] **************************************************",
                                "Friday 13 March 2026  01:23:36 +0000 (0:00:03.268)       0:00:03.320 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add manager role to user test-admin] *************************************",
                                "Friday 13 March 2026  01:23:39 +0000 (0:00:03.710)       0:00:07.030 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test project] *****************************************************",
                                "Friday 13 March 2026  01:23:46 +0000 (0:00:06.559)       0:00:13.590 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test user] ********************************************************",
                                "Friday 13 March 2026  01:23:50 +0000 (0:00:04.045)       0:00:17.635 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add member roles to user test] *******************************************",
                                "Friday 13 March 2026  01:23:54 +0000 (0:00:04.186)       0:00:21.822 ********** ",
                                "\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m",
                                "",
                                "TASK [Create test server group] ************************************************",
                                "Friday 13 March 2026  01:24:05 +0000 (0:00:11.380)       0:00:33.202 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create ssh security group] ***********************************************",
                                "Friday 13 March 2026  01:24:10 +0000 (0:00:04.789)       0:00:37.992 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to ssh security group] ******************************************",
                                "Friday 13 March 2026  01:24:14 +0000 (0:00:04.161)       0:00:42.153 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create icmp security group] **********************************************",
                                "Friday 13 March 2026  01:24:18 +0000 (0:00:03.831)       0:00:45.984 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to icmp security group] *****************************************",
                                "Friday 13 March 2026  01:24:22 +0000 (0:00:03.637)       0:00:49.622 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test keypair] *****************************************************",
                                "Friday 13 March 2026  01:24:26 +0000 (0:00:03.851)       0:00:53.474 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test network] *****************************************************",
                                "Friday 13 March 2026  01:24:29 +0000 (0:00:03.798)       0:00:57.273 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test subnet] ******************************************************",
                                "Friday 13 March 2026  01:24:34 +0000 (0:00:04.801)       0:01:02.074 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test router] ******************************************************",
                                "Friday 13 March 2026  01:24:40 +0000 (0:00:05.719)       0:01:07.794 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Manage test instances and volumes] ***************************************",
                                "",
                                "TASK [Get test server group] ***************************************************",
                                "Friday 13 March 2026  01:24:51 +0000 (0:00:10.649)       0:01:18.444 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Detach test volume] ******************************************************",
                                "Friday 13 March 2026  01:24:54 +0000 (0:00:03.414)       0:01:21.858 ********** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test volume] ******************************************************",
                                "Friday 13 March 2026  01:24:54 +0000 (0:00:00.053)       0:01:21.911 ********** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test instances] ***************************************************",
                                "Friday 13 March 2026  01:24:54 +0000 (0:00:00.039)       0:01:21.951 ********** ",
                                "\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Wait for instance deletion to complete] **********************************",
                                "Friday 13 March 2026  01:24:54 +0000 (0:00:00.158)       0:01:22.110 ********** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test instances] ***************************************************",
                                "Friday 13 March 2026  01:24:54 +0000 (0:00:00.138)       0:01:22.249 ********** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for instance creation to complete] **********************************",
                                "Friday 13 March 2026  01:24:59 +0000 (0:00:04.635)       0:01:26.885 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j430084601681.2616', 'results_file': '/ansible/.ansible_async/j430084601681.2616', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j504665637972.2641', 'results_file': '/ansible/.ansible_async/j504665637972.2641', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j984402392910.2666', 'results_file': '/ansible/.ansible_async/j984402392910.2666', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j508465868329.2691', 'results_file': '/ansible/.ansible_async/j508465868329.2691', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j65206343094.2716', 'results_file': '/ansible/.ansible_async/j65206343094.2716', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add metadata to instances] ***********************************************",
                                "Friday 13 March 2026  01:25:57 +0000 (0:00:57.655)       0:02:24.540 ********** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for metadata to be added] *******************************************",
                                "Friday 13 March 2026  01:26:01 +0000 (0:00:04.524)       0:02:29.065 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j154185062180.2827', 'results_file': '/ansible/.ansible_async/j154185062180.2827', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j218872645057.2852', 'results_file': '/ansible/.ansible_async/j218872645057.2852', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j220671646960.2877', 'results_file': '/ansible/.ansible_async/j220671646960.2877', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j915736299113.2902', 'results_file': '/ansible/.ansible_async/j915736299113.2902', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j294611988658.2927', 'results_file': '/ansible/.ansible_async/j294611988658.2927', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add tag to instances] ****************************************************",
                                "Friday 13 March 2026  01:26:11 +0000 (0:00:09.628)       0:02:38.693 ********** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for tags to be added] ***********************************************",
                                "Friday 13 March 2026  01:26:15 +0000 (0:00:04.182)       0:02:42.875 ********** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j966377388877.3003', 'results_file': '/ansible/.ansible_async/j966377388877.3003', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j927214934809.3028', 'results_file': '/ansible/.ansible_async/j927214934809.3028', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j964121494158.3054', 'results_file': '/ansible/.ansible_async/j964121494158.3054', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j287701880302.3080', 'results_file': '/ansible/.ansible_async/j287701880302.3080', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j272798000334.3106', 'results_file': '/ansible/.ansible_async/j272798000334.3106', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Create test volume] ******************************************************",
                                "Friday 13 March 2026  01:26:25 +0000 (0:00:09.488)       0:02:52.364 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Attach test volume] ******************************************************",
                                "Friday 13 March 2026  01:26:31 +0000 (0:00:06.542)       0:02:58.907 ********** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create floating ip address] **********************************************",
                                "Friday 13 March 2026  01:26:45 +0000 (0:00:13.454)       0:03:12.361 ********** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Print floating ip address] ***********************************************",
                                "Friday 13 March 2026  01:26:50 +0000 (0:00:04.999)       0:03:17.361 ********** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"192.168.112.161\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:26:50 +0000 (0:00:00.042)       0:03:17.403 ********** ",
                                "=============================================================================== ",
                                "Wait for instance creation to complete --------------------------------- 57.66s",
                                "Attach test volume ----------------------------------------------------- 13.45s",
                                "Add member roles to user test ------------------------------------------ 11.38s",
                                "Create test router ----------------------------------------------------- 10.65s",
                                "Wait for metadata to be added ------------------------------------------- 9.63s",
                                "Wait for tags to be added ----------------------------------------------- 9.49s",
                                "Add manager role to user test-admin ------------------------------------- 6.56s",
                                "Create test volume ------------------------------------------------------ 6.54s",
                                "Create test subnet ------------------------------------------------------ 5.72s",
                                "Create floating ip address ---------------------------------------------- 5.00s",
                                "Create test network ----------------------------------------------------- 4.80s",
                                "Create test server group ------------------------------------------------ 4.79s",
                                "Create test instances --------------------------------------------------- 4.64s",
                                "Add metadata to instances ----------------------------------------------- 4.53s",
                                "Create test user -------------------------------------------------------- 4.19s",
                                "Add tag to instances ---------------------------------------------------- 4.18s",
                                "Create ssh security group ----------------------------------------------- 4.16s",
                                "Create test project ----------------------------------------------------- 4.05s",
                                "Add rule to icmp security group ----------------------------------------- 3.85s",
                                "Add rule to ssh security group ------------------------------------------ 3.83s",
                                "+ server_list",
                                "+ openstack --os-cloud test server list",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE | test=192.168.112.141, 192.168.200.30  | N/A (booted from volume) | SCS-1L-1 |",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE | test=192.168.112.156, 192.168.200.80  | N/A (booted from volume) | SCS-1L-1 |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE | test=192.168.112.123, 192.168.200.121 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE | test=192.168.112.191, 192.168.200.13  | N/A (booted from volume) | SCS-1L-1 |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE | test=192.168.112.161, 192.168.200.236 | N/A (booted from volume) | SCS-1L-1 |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "+ openstack --os-cloud test server show test",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-13T01:25:30.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.161, 192.168.200.236                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-13T01:25:03Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | e6988b32-498d-46d2-b1fe-3cea242e8f3e                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-13T01:26:02Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='3857ac78-c588-4e7e-ba0a-80a8215a033b'                                                                                                                                                                                                                                                                                                             |",
                                "|                                     | delete_on_termination='False', id='30b2a964-3736-450d-84ca-79a85f4022cb'                                                                                                                                                                                                                                                                                                            |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-1",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-13T01:25:31.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.191, 192.168.200.13                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-13T01:25:03Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 46752925-b128-4628-8a08-dc074af15a37                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-13T01:26:03Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='2ade6197-02cf-46f9-bdd3-8730978e79f1'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-2",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-13T01:25:33.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.156, 192.168.200.80                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-13T01:25:05Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 488c8c86-dc7a-4826-9b1b-9abaf28c3d20                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='b36e9ede-44fb-447c-80dc-63b604527332'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-3",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-13T01:25:31.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.123, 192.168.200.121                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-13T01:25:05Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 70a1c14edc359cbcdf7297590131d983f923de78f2e9fc242c927e35                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 957c72c6-9401-4f8f-877a-087c45362881                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-13T01:26:04Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='db590097-db33-4136-8b60-1bc92af56e8e'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-4",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-03-13T01:25:34.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.141, 192.168.200.30                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-03-13T01:25:08Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='True', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | ea1c861b1c235aaadb8497ab6289fa1013a500fd3b8947c250900d9a                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 27cf0c91-b78c-4057-8905-5e463648a2a1                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | c7b30474d4ec441d830b893f4855fb71                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "|                                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-03-13T01:26:05Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | fd191c5428d74ff59eaf8492937373d7                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='6bfa19e9-07ae-4e96-9435-4be3397815d8'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ server_ping",
                                "++ tr -d '\\r'",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.123",
                                "PING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=5.69 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.36 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=2.07 ms",
                                "",
                                "--- 192.168.112.123 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 2.068/3.371/5.691/1.644 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.161",
                                "PING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=10.6 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=1.50 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.18 ms",
                                "",
                                "--- 192.168.112.161 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.178/4.422/10.586/4.360 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=4.26 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.18 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.93 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.934/2.792/4.259/1.042 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.44 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.45 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.69 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.694/3.194/5.441/1.618 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.191",
                                "PING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=4.89 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.18 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.80 ms",
                                "",
                                "--- 192.168.112.191 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.800/2.956/4.887/1.373 ms",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:24\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:28\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4",
                                "\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:27:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:28:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:49\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:52\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:29:56\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ tr -d '\\r'",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.123",
                                "PING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=4.71 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.52 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.74 ms",
                                "",
                                "--- 192.168.112.123 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.741/2.989/4.708/1.255 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.161",
                                "PING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=10.3 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.67 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.71 ms",
                                "",
                                "--- 192.168.112.161 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.705/4.906/10.349/3.868 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.18 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.24 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.53 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.527/3.979/8.175/2.981 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.07 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.60 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.71 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.707/3.126/5.074/1.424 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.191",
                                "PING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=6.09 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.05 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.85 ms",
                                "",
                                "--- 192.168.112.191 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.854/3.331/6.093/1.954 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5",
                                "\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo migratable instances found on node testbed-node-5\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.123",
                                "PING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=9.67 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.27 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.56 ms",
                                "",
                                "--- 192.168.112.123 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.561/4.501/9.668/3.665 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.161",
                                "PING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=6.14 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.72 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.72 ms",
                                "",
                                "--- 192.168.112.161 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.716/3.524/6.141/1.894 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=9.46 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=1.96 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.83 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.830/4.415/9.460/3.567 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.54 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.76 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.25 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.247/2.848/5.535/1.911 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.191",
                                "PING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=3.07 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.14 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.60 ms",
                                "",
                                "--- 192.168.112.191 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.598/2.270/3.073/0.609 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3",
                                "\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:38\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 27cf0c91-b78c-4057-8905-5e463648a2a1\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:30:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:31:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:32:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:18\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.123",
                                "PING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=5.53 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.54 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=1.66 ms",
                                "",
                                "--- 192.168.112.123 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.661/3.242/5.527/1.654 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.161",
                                "PING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=9.38 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=3.00 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.84 ms",
                                "",
                                "--- 192.168.112.161 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.841/4.741/9.379/3.313 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=8.82 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=1.92 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.97 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.922/4.238/8.820/3.239 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.29 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.77 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.68 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.681/2.915/5.290/1.679 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.191",
                                "PING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=4.94 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=1.97 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.53 ms",
                                "",
                                "--- 192.168.112.191 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.530/2.815/4.942/1.514 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4",
                                "\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:37\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 27cf0c91-b78c-4057-8905-5e463648a2a1\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:33:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 27cf0c91-b78c-4057-8905-5e463648a2a1 (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 488c8c86-dc7a-4826-9b1b-9abaf28c3d20\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 (test-2) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:34:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 957c72c6-9401-4f8f-877a-087c45362881\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 957c72c6-9401-4f8f-877a-087c45362881 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 46752925-b128-4628-8a08-dc074af15a37\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 46752925-b128-4628-8a08-dc074af15a37 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-03-13 01:35:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e6988b32-498d-46d2-b1fe-3cea242e8f3e\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e6988b32-498d-46d2-b1fe-3cea242e8f3e (test) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-03-13 01:36:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 27cf0c91-b78c-4057-8905-5e463648a2a1 | test-4 | ACTIVE   |",
                                "| 488c8c86-dc7a-4826-9b1b-9abaf28c3d20 | test-2 | ACTIVE   |",
                                "| 957c72c6-9401-4f8f-877a-087c45362881 | test-3 | ACTIVE   |",
                                "| 46752925-b128-4628-8a08-dc074af15a37 | test-1 | ACTIVE   |",
                                "| e6988b32-498d-46d2-b1fe-3cea242e8f3e | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.123",
                                "PING 192.168.112.123 (192.168.112.123) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.123: icmp_seq=1 ttl=63 time=9.30 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=2 ttl=63 time=2.28 ms",
                                "64 bytes from 192.168.112.123: icmp_seq=3 ttl=63 time=2.07 ms",
                                "",
                                "--- 192.168.112.123 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 2.069/4.549/9.297/3.358 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.161",
                                "PING 192.168.112.161 (192.168.112.161) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.161: icmp_seq=1 ttl=63 time=8.07 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=2 ttl=63 time=2.19 ms",
                                "64 bytes from 192.168.112.161: icmp_seq=3 ttl=63 time=1.75 ms",
                                "",
                                "--- 192.168.112.161 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.746/4.001/8.074/2.885 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.156",
                                "PING 192.168.112.156 (192.168.112.156) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.156: icmp_seq=1 ttl=63 time=4.63 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=2 ttl=63 time=2.12 ms",
                                "64 bytes from 192.168.112.156: icmp_seq=3 ttl=63 time=1.60 ms",
                                "",
                                "--- 192.168.112.156 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.604/2.784/4.634/1.324 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.77 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.75 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.70 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.697/3.071/5.767/1.906 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.191",
                                "PING 192.168.112.191 (192.168.112.191) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.191: icmp_seq=1 ttl=63 time=5.28 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=2 ttl=63 time=2.14 ms",
                                "64 bytes from 192.168.112.191: icmp_seq=3 ttl=63 time=1.74 ms",
                                "",
                                "--- 192.168.112.191 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.740/3.051/5.278/1.582 ms"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001e-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:36:47.291877Z",
                            "start": "2026-03-13T01:19:18.142886Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001e",
                        "name": "Run checks"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.46",
                                "/opt/configuration/scripts/check/302-openstack-with-tempest.sh"
                            ],
                            "delta": "0:01:55.582150",
                            "end": "2026-03-13 01:38:43.239752",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed03/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.46 /opt/configuration/scripts/check/302-openstack-with-tempest.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001f-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 01:36:47.657602",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# Tempest\n\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2025.1\n++ OPENSTACK_VERSION=2025.1\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# Tempest'\n+ echo\n+ [[ ! -e /opt/tempest ]]\n+ osism apply tempest --skip-tags run-tempest\n\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m\n\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d6bb3043-b58d-4b8b-ac13-37fe57882117 (tempest) was prepared for execution.\u001b[0m\n\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d6bb3043-b58d-4b8b-ac13-37fe57882117 (tempest) has been started and output is visible here.\u001b[0m\n\nPLAY [Run tempest] *************************************************************\n\nTASK [osism.validations.tempest : Create tempest workdir] **********************\nFriday 13 March 2026  01:37:04 +0000 (0:00:00.249)       0:00:00.249 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest wrapper script] *****************\nFriday 13 March 2026  01:37:05 +0000 (0:00:00.728)       0:00:00.977 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Check for existing tempest initialisation] ***\nFriday 13 March 2026  01:37:06 +0000 (0:00:01.259)       0:00:02.237 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Init tempest] ********************************\nFriday 13 March 2026  01:37:07 +0000 (0:00:00.411)       0:00:02.649 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve image IDs] ***************************\nFriday 13 March 2026  01:37:27 +0000 (0:00:20.384)       0:00:23.033 ********** \n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m\n\nTASK [osism.validations.tempest : Assert images have been resolved] ************\nFriday 13 March 2026  01:37:35 +0000 (0:00:08.198)       0:00:31.232 ********** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Get auth token] ******************************\nFriday 13 March 2026  01:37:35 +0000 (0:00:00.161)       0:00:31.394 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get endpoint catalog] ************************\nFriday 13 March 2026  01:37:39 +0000 (0:00:03.690)       0:00:35.085 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get service catalog] *************************\nFriday 13 March 2026  01:37:41 +0000 (0:00:01.880)       0:00:36.965 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Register img_file name] **********************\nFriday 13 March 2026  01:37:44 +0000 (0:00:03.552)       0:00:40.518 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Download img_file from image_ref] ************\nFriday 13 March 2026  01:37:45 +0000 (0:00:00.196)       0:00:40.714 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Install qemu-utils package] ******************\nFriday 13 March 2026  01:37:47 +0000 (0:00:02.388)       0:00:43.103 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Convert img_file to qcow2 format] ************\nFriday 13 March 2026  01:37:56 +0000 (0:00:09.073)       0:00:52.177 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get network API extensions] ******************\nFriday 13 March 2026  01:37:57 +0000 (0:00:00.766)       0:00:52.943 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Revoke token] ********************************\nFriday 13 March 2026  01:37:58 +0000 (0:00:01.616)       0:00:54.559 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option api_extensions] ***\nFriday 13 March 2026  01:38:00 +0000 (0:00:01.622)       0:00:56.182 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option img_file] *********\nFriday 13 March 2026  01:38:00 +0000 (0:00:00.208)       0:00:56.391 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve floating network ID] *****************\nFriday 13 March 2026  01:38:01 +0000 (0:00:00.199)       0:00:56.590 ********** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Assert floating network id has been resolved] ***\nFriday 13 March 2026  01:38:04 +0000 (0:00:03.888)       0:01:00.479 ********** \n\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Resolve flavor IDs] **************************\nFriday 13 March 2026  01:38:05 +0000 (0:00:00.198)       0:01:00.678 ********** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Assert flavors have been resolved] ***********\nFriday 13 March 2026  01:38:05 +0000 (0:00:00.384)       0:01:01.063 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of exclude list] *******************\nFriday 13 March 2026  01:38:05 +0000 (0:00:00.147)       0:01:01.211 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy exclude list] ***************************\nFriday 13 March 2026  01:38:06 +0000 (0:00:00.453)       0:01:01.664 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of include list] *******************\nFriday 13 March 2026  01:38:06 +0000 (0:00:00.878)       0:01:02.543 ********** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy include list] ***************************\nFriday 13 March 2026  01:38:07 +0000 (0:00:00.423)       0:01:02.967 ********** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Create tempest flavors] **********************\nFriday 13 March 2026  01:38:07 +0000 (0:00:00.140)       0:01:03.107 ********** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest.conf file] **********************\nFriday 13 March 2026  01:38:14 +0000 (0:00:07.358)       0:01:10.466 ********** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nFriday 13 March 2026  01:38:15 +0000 (0:00:01.043)       0:01:11.509 ********** \n=============================================================================== \nosism.validations.tempest : Init tempest ------------------------------- 20.38s\nosism.validations.tempest : Install qemu-utils package ------------------ 9.07s\nosism.validations.tempest : Resolve image IDs --------------------------- 8.20s\nosism.validations.tempest : Create tempest flavors ---------------------- 7.36s\nosism.validations.tempest : Resolve floating network ID ----------------- 3.89s\nosism.validations.tempest : Get auth token ------------------------------ 3.69s\nosism.validations.tempest : Get service catalog ------------------------- 3.55s\nosism.validations.tempest : Download img_file from image_ref ------------ 2.39s\nosism.validations.tempest : Get endpoint catalog ------------------------ 1.88s\nosism.validations.tempest : Revoke token -------------------------------- 1.62s\nosism.validations.tempest : Get network API extensions ------------------ 1.62s\nosism.validations.tempest : Copy tempest wrapper script ----------------- 1.26s\nosism.validations.tempest : Copy tempest.conf file ---------------------- 1.04s\nosism.validations.tempest : Copy exclude list --------------------------- 0.88s\nosism.validations.tempest : Convert img_file to qcow2 format ------------ 0.77s\nosism.validations.tempest : Create tempest workdir ---------------------- 0.73s\nosism.validations.tempest : Get stats of exclude list ------------------- 0.45s\nosism.validations.tempest : Get stats of include list ------------------- 0.42s\nosism.validations.tempest : Check for existing tempest initialisation --- 0.41s\nosism.validations.tempest : Resolve flavor IDs -------------------------- 0.38s\n+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf\n+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf\n+ [[ false == \\t\\r\\u\\e ]]\n\n## IDENTITY (API)\n\n+ echo\n+ echo '## IDENTITY (API)'\n+ echo\n+ _tempest tempest.api.identity.v3\n+ local regex=tempest.api.identity.v3\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## IMAGE (API)\n\n+ echo\n+ echo '## IMAGE (API)'\n+ echo\n+ _tempest tempest.api.image.v2\n+ local regex=tempest.api.image.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## NETWORK (API)\n\n+ echo\n+ echo '## NETWORK (API)'\n+ echo\n+ _tempest tempest.api.network\n+ local regex=tempest.api.network\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## VOLUME (API)\n\n+ echo\n+ echo '## VOLUME (API)'\n+ echo\n+ _tempest tempest.api.volume\n+ local regex=tempest.api.volume\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## COMPUTE (API)\n\n+ echo\n+ echo '## COMPUTE (API)'\n+ echo\n+ _tempest tempest.api.compute\n+ local regex=tempest.api.compute\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## DNS (API)\n\n+ echo\n+ echo '## DNS (API)'\n+ echo\n+ _tempest designate_tempest_plugin.tests.api.v2\n+ local regex=designate_tempest_plugin.tests.api.v2\n++ date +%Y%m%d-%H%M\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## OBJECT-STORE (API)\n\n+ echo\n+ echo '## OBJECT-STORE (API)'\n+ echo\n+ _tempest tempest.api.object_storage\n+ local regex=tempest.api.object_storage\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260313-0138.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init",
                            "stdout_lines": [
                                "",
                                "# Tempest",
                                "",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2025.1",
                                "++ OPENSTACK_VERSION=2025.1",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.46",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# Tempest'",
                                "+ echo",
                                "+ [[ ! -e /opt/tempest ]]",
                                "+ osism apply tempest --skip-tags run-tempest",
                                "\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m",
                                "\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask d6bb3043-b58d-4b8b-ac13-37fe57882117 (tempest) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-03-13 01:37:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task d6bb3043-b58d-4b8b-ac13-37fe57882117 (tempest) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Run tempest] *************************************************************",
                                "",
                                "TASK [osism.validations.tempest : Create tempest workdir] **********************",
                                "Friday 13 March 2026  01:37:04 +0000 (0:00:00.249)       0:00:00.249 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest wrapper script] *****************",
                                "Friday 13 March 2026  01:37:05 +0000 (0:00:00.728)       0:00:00.977 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Check for existing tempest initialisation] ***",
                                "Friday 13 March 2026  01:37:06 +0000 (0:00:01.259)       0:00:02.237 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Init tempest] ********************************",
                                "Friday 13 March 2026  01:37:07 +0000 (0:00:00.411)       0:00:02.649 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve image IDs] ***************************",
                                "Friday 13 March 2026  01:37:27 +0000 (0:00:20.384)       0:00:23.033 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert images have been resolved] ************",
                                "Friday 13 March 2026  01:37:35 +0000 (0:00:08.198)       0:00:31.232 ********** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get auth token] ******************************",
                                "Friday 13 March 2026  01:37:35 +0000 (0:00:00.161)       0:00:31.394 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get endpoint catalog] ************************",
                                "Friday 13 March 2026  01:37:39 +0000 (0:00:03.690)       0:00:35.085 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get service catalog] *************************",
                                "Friday 13 March 2026  01:37:41 +0000 (0:00:01.880)       0:00:36.965 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Register img_file name] **********************",
                                "Friday 13 March 2026  01:37:44 +0000 (0:00:03.552)       0:00:40.518 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Download img_file from image_ref] ************",
                                "Friday 13 March 2026  01:37:45 +0000 (0:00:00.196)       0:00:40.714 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Install qemu-utils package] ******************",
                                "Friday 13 March 2026  01:37:47 +0000 (0:00:02.388)       0:00:43.103 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Convert img_file to qcow2 format] ************",
                                "Friday 13 March 2026  01:37:56 +0000 (0:00:09.073)       0:00:52.177 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get network API extensions] ******************",
                                "Friday 13 March 2026  01:37:57 +0000 (0:00:00.766)       0:00:52.943 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Revoke token] ********************************",
                                "Friday 13 March 2026  01:37:58 +0000 (0:00:01.616)       0:00:54.559 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option api_extensions] ***",
                                "Friday 13 March 2026  01:38:00 +0000 (0:00:01.622)       0:00:56.182 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option img_file] *********",
                                "Friday 13 March 2026  01:38:00 +0000 (0:00:00.208)       0:00:56.391 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve floating network ID] *****************",
                                "Friday 13 March 2026  01:38:01 +0000 (0:00:00.199)       0:00:56.590 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert floating network id has been resolved] ***",
                                "Friday 13 March 2026  01:38:04 +0000 (0:00:03.888)       0:01:00.479 ********** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve flavor IDs] **************************",
                                "Friday 13 March 2026  01:38:05 +0000 (0:00:00.198)       0:01:00.678 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert flavors have been resolved] ***********",
                                "Friday 13 March 2026  01:38:05 +0000 (0:00:00.384)       0:01:01.063 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of exclude list] *******************",
                                "Friday 13 March 2026  01:38:05 +0000 (0:00:00.147)       0:01:01.211 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy exclude list] ***************************",
                                "Friday 13 March 2026  01:38:06 +0000 (0:00:00.453)       0:01:01.664 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of include list] *******************",
                                "Friday 13 March 2026  01:38:06 +0000 (0:00:00.878)       0:01:02.543 ********** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy include list] ***************************",
                                "Friday 13 March 2026  01:38:07 +0000 (0:00:00.423)       0:01:02.967 ********** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Create tempest flavors] **********************",
                                "Friday 13 March 2026  01:38:07 +0000 (0:00:00.140)       0:01:03.107 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest.conf file] **********************",
                                "Friday 13 March 2026  01:38:14 +0000 (0:00:07.358)       0:01:10.466 ********** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Friday 13 March 2026  01:38:15 +0000 (0:00:01.043)       0:01:11.509 ********** ",
                                "=============================================================================== ",
                                "osism.validations.tempest : Init tempest ------------------------------- 20.38s",
                                "osism.validations.tempest : Install qemu-utils package ------------------ 9.07s",
                                "osism.validations.tempest : Resolve image IDs --------------------------- 8.20s",
                                "osism.validations.tempest : Create tempest flavors ---------------------- 7.36s",
                                "osism.validations.tempest : Resolve floating network ID ----------------- 3.89s",
                                "osism.validations.tempest : Get auth token ------------------------------ 3.69s",
                                "osism.validations.tempest : Get service catalog ------------------------- 3.55s",
                                "osism.validations.tempest : Download img_file from image_ref ------------ 2.39s",
                                "osism.validations.tempest : Get endpoint catalog ------------------------ 1.88s",
                                "osism.validations.tempest : Revoke token -------------------------------- 1.62s",
                                "osism.validations.tempest : Get network API extensions ------------------ 1.62s",
                                "osism.validations.tempest : Copy tempest wrapper script ----------------- 1.26s",
                                "osism.validations.tempest : Copy tempest.conf file ---------------------- 1.04s",
                                "osism.validations.tempest : Copy exclude list --------------------------- 0.88s",
                                "osism.validations.tempest : Convert img_file to qcow2 format ------------ 0.77s",
                                "osism.validations.tempest : Create tempest workdir ---------------------- 0.73s",
                                "osism.validations.tempest : Get stats of exclude list ------------------- 0.45s",
                                "osism.validations.tempest : Get stats of include list ------------------- 0.42s",
                                "osism.validations.tempest : Check for existing tempest initialisation --- 0.41s",
                                "osism.validations.tempest : Resolve flavor IDs -------------------------- 0.38s",
                                "+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf",
                                "+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf",
                                "+ [[ false == \\t\\r\\u\\e ]]",
                                "",
                                "## IDENTITY (API)",
                                "",
                                "+ echo",
                                "+ echo '## IDENTITY (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.identity.v3",
                                "+ local regex=tempest.api.identity.v3",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## IMAGE (API)",
                                "",
                                "+ echo",
                                "+ echo '## IMAGE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.image.v2",
                                "+ local regex=tempest.api.image.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## NETWORK (API)",
                                "",
                                "+ echo",
                                "+ echo '## NETWORK (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.network",
                                "+ local regex=tempest.api.network",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## VOLUME (API)",
                                "",
                                "+ echo",
                                "+ echo '## VOLUME (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.volume",
                                "+ local regex=tempest.api.volume",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## COMPUTE (API)",
                                "",
                                "+ echo",
                                "+ echo '## COMPUTE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.compute",
                                "+ local regex=tempest.api.compute",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## DNS (API)",
                                "",
                                "+ echo",
                                "+ echo '## DNS (API)'",
                                "+ echo",
                                "+ _tempest designate_tempest_plugin.tests.api.v2",
                                "+ local regex=designate_tempest_plugin.tests.api.v2",
                                "++ date +%Y%m%d-%H%M",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## OBJECT-STORE (API)",
                                "",
                                "+ echo",
                                "+ echo '## OBJECT-STORE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.object_storage",
                                "+ local regex=tempest.api.object_storage",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260313-0138.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init"
                            ],
                            "zuul_log_id": "0242ac1f-6606-6cb6-5436-00000000001f-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:43.447155Z",
                            "start": "2026-03-13T01:36:47.344900Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-00000000001f",
                        "name": "Run tempest"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_prometheus_alert_status | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:44.001374Z",
                            "start": "2026-03-13T01:38:43.490017Z"
                        },
                        "id": "0242ac1f-6606-6cb6-5436-000000000020",
                        "name": "Check prometheus alert status"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 12,
            "failures": 0,
            "ignored": 0,
            "ok": 25,
            "rescued": 0,
            "skipped": 4,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:38:47.861055Z",
                    "start": "2026-03-13T01:38:45.014608Z"
                },
                "id": "0242ac1f-6606-12fe-a39e-000000000002",
                "name": "Post output play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:45.080575Z",
                            "start": "2026-03-13T01:38:45.033887Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000007",
                        "name": "Register sources"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "sudo",
                                "-n",
                                "true"
                            ],
                            "delta": "0:00:00.016550",
                            "end": "2026-03-13 01:38:45.983836",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "sudo -n true",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-12fe-a39e-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-03-13 01:38:45.967286",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sudo: a password is required",
                            "stdout_lines": [
                                "sudo: a password is required"
                            ],
                            "zuul_log_id": "0242ac1f-6606-12fe-a39e-000000000009-1-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:46.135141Z",
                            "start": "2026-03-13T01:38:45.122878Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000009",
                        "name": "Check sudo"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:46.176490Z",
                            "start": "2026-03-13T01:38:46.150499Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000000a",
                        "name": "Set source and destination for files and folders"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "all_sources": []
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:46.263474Z",
                            "start": "2026-03-13T01:38:46.193273Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000000c",
                        "name": "Build a list of source, dest dictionaries"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/docs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/docs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "docs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed03/zuul-output/logs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed03/zuul-output/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1004,
                                    "group": "zuul-testbed03",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed03",
                                            "path": "/home/zuul-testbed03/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed03",
                                    "path": "/home/zuul-testbed03/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1003,
                                    "zj_output_dirs": "logs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.344119Z",
                            "start": "2026-03-13T01:38:46.271596Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000000d",
                        "name": "Ensure target folders exist"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "results": [],
                            "skipped": true,
                            "skipped_reason": "No items in the list"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.379877Z",
                            "start": "2026-03-13T01:38:47.354780Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000000f",
                        "name": "Copy files and folders to staging folder"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed03/zuul-output/logs"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed03/zuul-output/logs"
                                }
                            },
                            "gid": 1004,
                            "group": "zuul-testbed03",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "dest": "/home/zuul-testbed03/zuul-output/logs",
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": "u=rwX,g=rX,o=rX",
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed03/zuul-output/logs",
                                    "recurse": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed03",
                            "path": "/home/zuul-testbed03/zuul-output/logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1003
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.739042Z",
                            "start": "2026-03-13T01:38:47.395142Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000011",
                        "name": "Make all log files readable"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stage_output_renames",
                            "changed": false,
                            "false_condition": "extensions_to_txt is defined and extensions_to_txt is not none",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.782764Z",
                            "start": "2026-03-13T01:38:47.747276Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000012",
                        "name": "Rename log files that match extensions_to_txt"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "find",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.824091Z",
                            "start": "2026-03-13T01:38:47.798459Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000014",
                        "name": "Discover log files for compression"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-12fe-a39e-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.861055Z",
                            "start": "2026-03-13T01:38:47.836322Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000015",
                        "name": "Archive everything from logs"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:39:09.548818Z",
                    "start": "2026-03-13T01:38:47.877595Z"
                },
                "id": "0242ac1f-6606-12fe-a39e-000000000017",
                "name": "Post cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.934968Z",
                            "start": "2026-03-13T01:38:47.885731Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-000000000019",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:38:47.972124Z",
                            "start": "2026-03-13T01:38:47.947593Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000001a",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:20.792489",
                            "end": "2026-03-13 01:39:09.092334",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-12fe-a39e-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 01:38:48.299845",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-13 01:38:49 - clean up servers\n2026-03-13 01:38:50 - testbed-manager\n2026-03-13 01:38:50 - testbed-node-0\n2026-03-13 01:38:50 - testbed-node-2\n2026-03-13 01:38:50 - testbed-node-5\n2026-03-13 01:38:50 - testbed-node-1\n2026-03-13 01:38:50 - testbed-node-3\n2026-03-13 01:38:50 - testbed-node-4\n2026-03-13 01:38:51 - clean up keypairs\n2026-03-13 01:38:51 - testbed\n2026-03-13 01:38:51 - wait for servers to be gone\n2026-03-13 01:39:01 - clean up ports\n2026-03-13 01:39:02 - 0e20cc2d-f5c4-43e4-bf07-70cac2a883f1\n2026-03-13 01:39:02 - 38bd7b83-ef2a-4fd5-b9d7-0ddfaf83a6de\n2026-03-13 01:39:02 - 670071ab-2c27-48c5-a098-e705a52445bb\n2026-03-13 01:39:03 - 91da1218-c10d-476c-bd3a-07c5008fc822\n2026-03-13 01:39:03 - b05cadca-a1ba-40c8-8416-3d17d0ec1cf4\n2026-03-13 01:39:03 - c7ea0029-231a-461d-987b-fe01ea7248e2\n2026-03-13 01:39:03 - cd46a55b-3127-464a-845c-fe24331a784d\n2026-03-13 01:39:03 - clean up volumes\n2026-03-13 01:39:04 - testbed-volume-3-node-base\n2026-03-13 01:39:04 - testbed-volume-1-node-base\n2026-03-13 01:39:04 - testbed-volume-0-node-base\n2026-03-13 01:39:04 - testbed-volume-5-node-base\n2026-03-13 01:39:04 - testbed-volume-4-node-base\n2026-03-13 01:39:04 - testbed-volume-2-node-base\n2026-03-13 01:39:04 - testbed-volume-manager-base\n2026-03-13 01:39:04 - testbed-volume-0-node-3\n2026-03-13 01:39:04 - testbed-volume-1-node-4\n2026-03-13 01:39:04 - testbed-volume-4-node-4\n2026-03-13 01:39:04 - testbed-volume-8-node-5\n2026-03-13 01:39:04 - testbed-volume-2-node-5\n2026-03-13 01:39:04 - testbed-volume-5-node-5\n2026-03-13 01:39:04 - testbed-volume-7-node-4\n2026-03-13 01:39:04 - testbed-volume-3-node-3\n2026-03-13 01:39:04 - testbed-volume-6-node-3\n2026-03-13 01:39:04 - disconnect routers\n2026-03-13 01:39:04 - testbed\n2026-03-13 01:39:05 - clean up subnets\n2026-03-13 01:39:05 - subnet-testbed-management\n2026-03-13 01:39:06 - clean up networks\n2026-03-13 01:39:06 - net-testbed-management\n2026-03-13 01:39:06 - clean up security groups\n2026-03-13 01:39:07 - testbed-management\n2026-03-13 01:39:07 - testbed-node\n2026-03-13 01:39:07 - clean up floating ips\n2026-03-13 01:39:07 - 81.163.192.46\n2026-03-13 01:39:07 - clean up routers\n2026-03-13 01:39:07 - testbed",
                            "stdout_lines": [
                                "2026-03-13 01:38:49 - clean up servers",
                                "2026-03-13 01:38:50 - testbed-manager",
                                "2026-03-13 01:38:50 - testbed-node-0",
                                "2026-03-13 01:38:50 - testbed-node-2",
                                "2026-03-13 01:38:50 - testbed-node-5",
                                "2026-03-13 01:38:50 - testbed-node-1",
                                "2026-03-13 01:38:50 - testbed-node-3",
                                "2026-03-13 01:38:50 - testbed-node-4",
                                "2026-03-13 01:38:51 - clean up keypairs",
                                "2026-03-13 01:38:51 - testbed",
                                "2026-03-13 01:38:51 - wait for servers to be gone",
                                "2026-03-13 01:39:01 - clean up ports",
                                "2026-03-13 01:39:02 - 0e20cc2d-f5c4-43e4-bf07-70cac2a883f1",
                                "2026-03-13 01:39:02 - 38bd7b83-ef2a-4fd5-b9d7-0ddfaf83a6de",
                                "2026-03-13 01:39:02 - 670071ab-2c27-48c5-a098-e705a52445bb",
                                "2026-03-13 01:39:03 - 91da1218-c10d-476c-bd3a-07c5008fc822",
                                "2026-03-13 01:39:03 - b05cadca-a1ba-40c8-8416-3d17d0ec1cf4",
                                "2026-03-13 01:39:03 - c7ea0029-231a-461d-987b-fe01ea7248e2",
                                "2026-03-13 01:39:03 - cd46a55b-3127-464a-845c-fe24331a784d",
                                "2026-03-13 01:39:03 - clean up volumes",
                                "2026-03-13 01:39:04 - testbed-volume-3-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-1-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-0-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-5-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-4-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-2-node-base",
                                "2026-03-13 01:39:04 - testbed-volume-manager-base",
                                "2026-03-13 01:39:04 - testbed-volume-0-node-3",
                                "2026-03-13 01:39:04 - testbed-volume-1-node-4",
                                "2026-03-13 01:39:04 - testbed-volume-4-node-4",
                                "2026-03-13 01:39:04 - testbed-volume-8-node-5",
                                "2026-03-13 01:39:04 - testbed-volume-2-node-5",
                                "2026-03-13 01:39:04 - testbed-volume-5-node-5",
                                "2026-03-13 01:39:04 - testbed-volume-7-node-4",
                                "2026-03-13 01:39:04 - testbed-volume-3-node-3",
                                "2026-03-13 01:39:04 - testbed-volume-6-node-3",
                                "2026-03-13 01:39:04 - disconnect routers",
                                "2026-03-13 01:39:04 - testbed",
                                "2026-03-13 01:39:05 - clean up subnets",
                                "2026-03-13 01:39:05 - subnet-testbed-management",
                                "2026-03-13 01:39:06 - clean up networks",
                                "2026-03-13 01:39:06 - net-testbed-management",
                                "2026-03-13 01:39:06 - clean up security groups",
                                "2026-03-13 01:39:07 - testbed-management",
                                "2026-03-13 01:39:07 - testbed-node",
                                "2026-03-13 01:39:07 - clean up floating ips",
                                "2026-03-13 01:39:07 - 81.163.192.46",
                                "2026-03-13 01:39:07 - clean up routers",
                                "2026-03-13 01:39:07 - testbed"
                            ],
                            "zuul_log_id": "0242ac1f-6606-12fe-a39e-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:09.548818Z",
                            "start": "2026-03-13T01:38:48.011684Z"
                        },
                        "id": "0242ac1f-6606-12fe-a39e-00000000001b",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 7,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/cleanup.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:39:13.206505Z",
                    "start": "2026-03-13T01:39:10.526502Z"
                },
                "id": "0242ac1f-6606-e5c1-54d7-000000000002",
                "name": "Cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:10.599060Z",
                            "start": "2026-03-13T01:39:10.544063Z"
                        },
                        "id": "0242ac1f-6606-e5c1-54d7-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:10.643400Z",
                            "start": "2026-03-13T01:39:10.608201Z"
                        },
                        "id": "0242ac1f-6606-e5c1-54d7-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.466020",
                            "end": "2026-03-13 01:39:13.005313",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed03/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-e5c1-54d7-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-03-13 01:39:11.539293",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-03-13 01:39:11 - clean up servers\n2026-03-13 01:39:12 - clean up keypairs\n2026-03-13 01:39:12 - wait for servers to be gone\n2026-03-13 01:39:12 - clean up ports\n2026-03-13 01:39:12 - clean up volumes\n2026-03-13 01:39:12 - disconnect routers\n2026-03-13 01:39:12 - clean up subnets\n2026-03-13 01:39:12 - clean up networks\n2026-03-13 01:39:12 - clean up security groups\n2026-03-13 01:39:12 - clean up floating ips\n2026-03-13 01:39:12 - clean up routers",
                            "stdout_lines": [
                                "2026-03-13 01:39:11 - clean up servers",
                                "2026-03-13 01:39:12 - clean up keypairs",
                                "2026-03-13 01:39:12 - wait for servers to be gone",
                                "2026-03-13 01:39:12 - clean up ports",
                                "2026-03-13 01:39:12 - clean up volumes",
                                "2026-03-13 01:39:12 - disconnect routers",
                                "2026-03-13 01:39:12 - clean up subnets",
                                "2026-03-13 01:39:12 - clean up networks",
                                "2026-03-13 01:39:12 - clean up security groups",
                                "2026-03-13 01:39:12 - clean up floating ips",
                                "2026-03-13 01:39:12 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-e5c1-54d7-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:13.206505Z",
                            "start": "2026-03-13T01:39:10.692348Z"
                        },
                        "id": "0242ac1f-6606-e5c1-54d7-000000000006",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "2",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post-fetch.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:39:17.753488Z",
                    "start": "2026-03-13T01:39:14.147905Z"
                },
                "id": "0242ac1f-6606-7baa-28bb-000000000002",
                "name": "Base post-fetch"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7baa-28bb-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:14.220072Z",
                            "start": "2026-03-13T01:39:14.163566Z"
                        },
                        "id": "0242ac1f-6606-7baa-28bb-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7baa-28bb-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:14.285544Z",
                            "start": "2026-03-13T01:39:14.234808Z"
                        },
                        "id": "0242ac1f-6606-7baa-28bb-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7baa-28bb-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:15.320382Z",
                            "start": "2026-03-13T01:39:14.294443Z"
                        },
                        "id": "0242ac1f-6606-7baa-28bb-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/logs/ /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/artifacts/ /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/docs/ /var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed03@81.163.193.140:/home/zuul-testbed03/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7baa-28bb-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:17.695226Z",
                            "start": "2026-03-13T01:39:15.334255Z"
                        },
                        "id": "0242ac1f-6606-7baa-28bb-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "msg": "All items skipped",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "docs"
                                }
                            ],
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-7baa-28bb-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:17.753488Z",
                            "start": "2026-03-13T01:39:17.707805Z"
                        },
                        "id": "0242ac1f-6606-7baa-28bb-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 3,
            "rescued": 0,
            "skipped": 2,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "3",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-03-13T01:39:19.833127Z",
                    "start": "2026-03-13T01:39:18.668647Z"
                },
                "id": "0242ac1f-6606-f9be-e3b1-000000000002",
                "name": "Base post"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC9hZOK4zg7iDGqaFxQ9uy0ic4PLr3V1cgsb+AkMti0USqK3vbO/cXhafoyXykJhOxCykz2WfdbDKSFitk67gUPtm5U6DRz+wLcL4t9MZubLLDYqucs7UjBEzky8PUsAdyhZqp/caf+M+mep9zOhstPwVHwpfxXJFieFSGC32gOnkTeUSFJE1CxcdEFnFsd+ofs5hKIXNRxRFivaIxkqemyTix6oabXVeY3dn/Qw1blXMMbNyt94M1R1REx1nREXvX8gQNLUx4pDYNCRE+q7CgRkCNN4CSkP+g+ZhjAU80p0DyQh0Xje4du/EzBPEd8dMj9VvgrXLOg3qzVIXOvT9fLZUooH19iT7qvxQuj6FI393nssuR9/qUMmTTKIeb4YnY1Y8Mg7YQNO+eVTzbizPQiJSd+2Dk1K21/H/Md54lHS6ZhUKzbuap/nl62MwdWAbqMexqXl9Vlb1CW7yDWVEpDArPEwSWyOAsZkqcE3W3re8xlD1pqPAqtivXAgzqFYxM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul-testbed03",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC9hZOK4zg7iDGqaFxQ9uy0ic4PLr3V1cgsb+AkMti0USqK3vbO/cXhafoyXykJhOxCykz2WfdbDKSFitk67gUPtm5U6DRz+wLcL4t9MZubLLDYqucs7UjBEzky8PUsAdyhZqp/caf+M+mep9zOhstPwVHwpfxXJFieFSGC32gOnkTeUSFJE1CxcdEFnFsd+ofs5hKIXNRxRFivaIxkqemyTix6oabXVeY3dn/Qw1blXMMbNyt94M1R1REx1nREXvX8gQNLUx4pDYNCRE+q7CgRkCNN4CSkP+g+ZhjAU80p0DyQh0Xje4du/EzBPEd8dMj9VvgrXLOg3qzVIXOvT9fLZUooH19iT7qvxQuj6FI393nssuR9/qUMmTTKIeb4YnY1Y8Mg7YQNO+eVTzbizPQiJSd+2Dk1K21/H/Md54lHS6ZhUKzbuap/nl62MwdWAbqMexqXl9Vlb1CW7yDWVEpDArPEwSWyOAsZkqcE3W3re8xlD1pqPAqtivXAgzqFYxM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed03/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul-testbed03",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-f9be-e3b1-000000000005",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/14c2d5479cea4ebc868c9c5d14dcb57c/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-03-13T01:39:19.833127Z",
                            "start": "2026-03-13T01:39:18.684345Z"
                        },
                        "id": "0242ac1f-6606-f9be-e3b1-000000000007",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
